var/home/core/zuul-output/0000755000175000017500000000000015067422757014544 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067440257015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005342507215067440250017710 0ustar rootrootOct 02 07:16:47 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 07:16:47 crc restorecon[4669]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:47 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 07:16:48 crc restorecon[4669]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 07:16:49 crc kubenswrapper[4829]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:49 crc kubenswrapper[4829]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 07:16:49 crc kubenswrapper[4829]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:49 crc kubenswrapper[4829]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:49 crc kubenswrapper[4829]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 07:16:49 crc kubenswrapper[4829]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.188021 4829 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193151 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193183 4829 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193192 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193202 4829 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193214 4829 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193252 4829 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193262 4829 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193270 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193280 4829 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193288 4829 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193296 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193304 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193312 4829 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193331 4829 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193340 4829 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193348 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193356 4829 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193363 4829 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193371 4829 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193379 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193389 4829 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193400 4829 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193408 4829 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193418 4829 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193427 4829 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193436 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193445 4829 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193454 4829 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193464 4829 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193474 4829 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193483 4829 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193492 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193502 4829 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193511 4829 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193520 4829 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193529 4829 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193537 4829 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193545 4829 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193554 4829 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193562 4829 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193570 4829 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193578 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193585 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193593 4829 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193603 4829 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193615 4829 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193625 4829 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193633 4829 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193642 4829 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193651 4829 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193660 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193668 4829 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193676 4829 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193684 4829 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193693 4829 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193701 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193710 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193718 4829 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193727 4829 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193737 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193745 4829 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193753 4829 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193761 4829 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193769 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193777 4829 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193785 4829 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193793 4829 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193800 4829 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193809 4829 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193817 4829 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.193824 4829 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.193969 4829 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.193985 4829 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.193999 4829 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194010 4829 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194021 4829 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194031 4829 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194043 4829 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194054 4829 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194063 4829 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194072 4829 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194083 4829 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194093 4829 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194102 4829 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194112 4829 flags.go:64] FLAG: --cgroup-root="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194121 4829 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194130 4829 flags.go:64] FLAG: --client-ca-file="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194139 4829 flags.go:64] FLAG: --cloud-config="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194148 4829 flags.go:64] FLAG: --cloud-provider="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194156 4829 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194168 4829 flags.go:64] FLAG: --cluster-domain="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194177 4829 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194187 4829 flags.go:64] FLAG: --config-dir="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194196 4829 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194205 4829 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194216 4829 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194250 4829 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194260 4829 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194269 4829 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194278 4829 flags.go:64] FLAG: --contention-profiling="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194286 4829 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194295 4829 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194305 4829 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194314 4829 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194324 4829 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194333 4829 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194342 4829 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194351 4829 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194360 4829 flags.go:64] FLAG: --enable-server="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194368 4829 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194382 4829 flags.go:64] FLAG: --event-burst="100" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194391 4829 flags.go:64] FLAG: --event-qps="50" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194400 4829 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194410 4829 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194418 4829 flags.go:64] FLAG: --eviction-hard="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194428 4829 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194437 4829 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194448 4829 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194460 4829 flags.go:64] FLAG: --eviction-soft="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194470 4829 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194479 4829 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194488 4829 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194516 4829 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194525 4829 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194534 4829 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194542 4829 flags.go:64] FLAG: --feature-gates="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194553 4829 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194563 4829 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194575 4829 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194587 4829 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194598 4829 flags.go:64] FLAG: --healthz-port="10248" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194607 4829 flags.go:64] FLAG: --help="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194616 4829 flags.go:64] FLAG: --hostname-override="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194625 4829 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194635 4829 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194645 4829 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194653 4829 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194663 4829 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194672 4829 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194681 4829 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194689 4829 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194698 4829 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194707 4829 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194716 4829 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194725 4829 flags.go:64] FLAG: --kube-reserved="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194734 4829 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194742 4829 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194751 4829 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194760 4829 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194769 4829 flags.go:64] FLAG: --lock-file="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194778 4829 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194787 4829 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194796 4829 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194809 4829 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194818 4829 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194827 4829 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194836 4829 flags.go:64] FLAG: --logging-format="text" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194845 4829 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194854 4829 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194862 4829 flags.go:64] FLAG: --manifest-url="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194872 4829 flags.go:64] FLAG: --manifest-url-header="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194890 4829 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194898 4829 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194909 4829 flags.go:64] FLAG: --max-pods="110" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194918 4829 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194927 4829 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194936 4829 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194945 4829 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194954 4829 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194963 4829 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194972 4829 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194990 4829 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.194999 4829 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195008 4829 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195017 4829 flags.go:64] FLAG: --pod-cidr="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195026 4829 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195038 4829 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195047 4829 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195056 4829 flags.go:64] FLAG: --pods-per-core="0" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195065 4829 flags.go:64] FLAG: --port="10250" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195075 4829 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195084 4829 flags.go:64] FLAG: --provider-id="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195093 4829 flags.go:64] FLAG: --qos-reserved="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195102 4829 flags.go:64] FLAG: --read-only-port="10255" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195110 4829 flags.go:64] FLAG: --register-node="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195119 4829 flags.go:64] FLAG: --register-schedulable="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195128 4829 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195142 4829 flags.go:64] FLAG: --registry-burst="10" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195151 4829 flags.go:64] FLAG: --registry-qps="5" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195160 4829 flags.go:64] FLAG: --reserved-cpus="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195170 4829 flags.go:64] FLAG: --reserved-memory="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195181 4829 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195197 4829 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195207 4829 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195244 4829 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195256 4829 flags.go:64] FLAG: --runonce="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195266 4829 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195277 4829 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195288 4829 flags.go:64] FLAG: --seccomp-default="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195298 4829 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195308 4829 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195319 4829 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195329 4829 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195340 4829 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195350 4829 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195360 4829 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195371 4829 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195381 4829 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195390 4829 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195399 4829 flags.go:64] FLAG: --system-cgroups="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195408 4829 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195422 4829 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195431 4829 flags.go:64] FLAG: --tls-cert-file="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195440 4829 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195451 4829 flags.go:64] FLAG: --tls-min-version="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195459 4829 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195468 4829 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195477 4829 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195486 4829 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195494 4829 flags.go:64] FLAG: --v="2" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195505 4829 flags.go:64] FLAG: --version="false" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195516 4829 flags.go:64] FLAG: --vmodule="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195527 4829 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.195536 4829 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195734 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195745 4829 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195755 4829 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195765 4829 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195773 4829 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195782 4829 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195789 4829 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195797 4829 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195805 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195813 4829 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195821 4829 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195828 4829 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195836 4829 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195844 4829 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195851 4829 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195859 4829 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195866 4829 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195874 4829 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195881 4829 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195896 4829 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195903 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195911 4829 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195919 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195926 4829 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195934 4829 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195942 4829 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195949 4829 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195957 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195965 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195975 4829 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195984 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.195992 4829 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196001 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196009 4829 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196017 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196025 4829 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196033 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196043 4829 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196051 4829 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196059 4829 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196067 4829 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196079 4829 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196087 4829 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196097 4829 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196107 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196118 4829 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196128 4829 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196137 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196145 4829 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196156 4829 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196165 4829 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196176 4829 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196219 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196249 4829 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196257 4829 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196267 4829 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196275 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196284 4829 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196293 4829 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196303 4829 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196311 4829 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196320 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196327 4829 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196336 4829 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196345 4829 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196353 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196362 4829 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196370 4829 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196378 4829 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196386 4829 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.196394 4829 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.196417 4829 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.209801 4829 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.210160 4829 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210358 4829 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210374 4829 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210383 4829 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210392 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210402 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210410 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210420 4829 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210428 4829 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210436 4829 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210444 4829 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210452 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210460 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210468 4829 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210476 4829 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210484 4829 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210491 4829 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210500 4829 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210511 4829 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210522 4829 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210531 4829 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210541 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210550 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210559 4829 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210568 4829 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210576 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210584 4829 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210596 4829 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210609 4829 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210618 4829 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210644 4829 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210652 4829 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210660 4829 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210668 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210676 4829 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210684 4829 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210691 4829 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210699 4829 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210708 4829 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210716 4829 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210723 4829 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210731 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210739 4829 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210747 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210755 4829 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210763 4829 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210771 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210780 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210787 4829 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210795 4829 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210806 4829 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210814 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210822 4829 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210830 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210838 4829 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210846 4829 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210859 4829 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210870 4829 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210879 4829 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210887 4829 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210894 4829 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210902 4829 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210910 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210917 4829 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210925 4829 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210933 4829 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210941 4829 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210948 4829 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210956 4829 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210964 4829 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210971 4829 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.210980 4829 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.210994 4829 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211216 4829 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211250 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211259 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211270 4829 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211281 4829 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211290 4829 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211298 4829 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211307 4829 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211315 4829 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211322 4829 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211330 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211338 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211345 4829 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211353 4829 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211361 4829 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211368 4829 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211376 4829 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211384 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211392 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211400 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211411 4829 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211420 4829 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211429 4829 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211438 4829 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211446 4829 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211454 4829 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211464 4829 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211474 4829 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211482 4829 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211491 4829 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211499 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211508 4829 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211518 4829 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211526 4829 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211534 4829 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211542 4829 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211550 4829 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211558 4829 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211565 4829 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211573 4829 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211581 4829 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211589 4829 feature_gate.go:330] unrecognized feature gate: Example Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211596 4829 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211604 4829 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211612 4829 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211620 4829 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211628 4829 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211636 4829 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211644 4829 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211652 4829 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211659 4829 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211667 4829 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211677 4829 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211687 4829 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211695 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211703 4829 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211712 4829 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211720 4829 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211727 4829 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211736 4829 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211744 4829 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211752 4829 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211759 4829 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211767 4829 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211776 4829 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211783 4829 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211791 4829 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211798 4829 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211806 4829 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211814 4829 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.211821 4829 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.211834 4829 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.212091 4829 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.220331 4829 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.220468 4829 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.223686 4829 server.go:997] "Starting client certificate rotation" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.223734 4829 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.223902 4829 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-22 16:24:42.010397641 +0000 UTC Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.224000 4829 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1233h7m52.786402762s for next certificate rotation Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.255022 4829 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.258313 4829 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.288209 4829 log.go:25] "Validated CRI v1 runtime API" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.324651 4829 log.go:25] "Validated CRI v1 image API" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.327453 4829 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.336962 4829 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-07-11-40-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.337015 4829 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.357473 4829 manager.go:217] Machine: {Timestamp:2025-10-02 07:16:49.355610363 +0000 UTC m=+0.695258788 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799886 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:95b96f34-f875-43c2-b6d4-5079b80b707a BootID:66ad52c8-80dd-4f56-a08f-df32feb03baf Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:fb:14:33 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:fb:14:33 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:cf:c0:97 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e9:da:f1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:29:88:d3 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:bc:54:f4 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:52:06:1c:17:25 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:6a:69:4f:07:0c:79 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.357824 4829 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.357980 4829 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.360428 4829 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.360608 4829 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.360647 4829 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.360871 4829 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.360882 4829 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.361463 4829 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.361508 4829 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.361707 4829 state_mem.go:36] "Initialized new in-memory state store" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.361811 4829 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.367506 4829 kubelet.go:418] "Attempting to sync node with API server" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.367532 4829 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.367561 4829 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.367575 4829 kubelet.go:324] "Adding apiserver pod source" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.367587 4829 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.371928 4829 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.372710 4829 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.377420 4829 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.379589 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.379644 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.379691 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379612 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379842 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379860 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379874 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.379792 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379897 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379911 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379924 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379948 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.379989 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.380011 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.380043 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.380058 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.381067 4829 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.383897 4829 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.384015 4829 server.go:1280] "Started kubelet" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.384283 4829 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.385548 4829 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.385989 4829 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.386016 4829 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.386057 4829 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 23:04:04.395171113 +0000 UTC Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.386084 4829 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1575h47m15.009088464s for next certificate rotation Oct 02 07:16:49 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.386573 4829 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.386583 4829 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.386681 4829 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.386971 4829 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.387739 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.387902 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.388213 4829 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="200ms" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.388418 4829 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.388861 4829 factory.go:55] Registering systemd factory Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.388923 4829 factory.go:221] Registration of the systemd container factory successfully Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.389408 4829 factory.go:153] Registering CRI-O factory Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.389452 4829 factory.go:221] Registration of the crio container factory successfully Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.389565 4829 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.389597 4829 factory.go:103] Registering Raw factory Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.389620 4829 manager.go:1196] Started watching for new ooms in manager Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.391066 4829 server.go:460] "Adding debug handlers to kubelet server" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.391074 4829 manager.go:319] Starting recovery of all containers Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.393453 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394539 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394746 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394800 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394823 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394859 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394880 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394895 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394924 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394940 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394962 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394978 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.394998 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395019 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395040 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395055 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395072 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395088 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395104 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395126 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395143 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395165 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395179 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395194 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395214 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395285 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395309 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395331 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395345 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395472 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.395498 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400346 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400422 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400462 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400482 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400513 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400534 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400553 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400579 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400600 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400622 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400649 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400670 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400696 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400719 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400741 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400770 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400800 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400829 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400849 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400868 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400894 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400934 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400958 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.400986 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401019 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401043 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401071 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401090 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401110 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401139 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401206 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401289 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401310 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401361 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401388 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401408 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401428 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401454 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401476 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401501 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401520 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401540 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401657 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401677 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401701 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401721 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401744 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401770 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401789 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401815 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401836 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401855 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401880 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401901 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401926 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401947 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401965 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.401990 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402013 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402034 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402060 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402089 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402118 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402139 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402159 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402189 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402212 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402263 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402286 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402308 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402335 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402355 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402381 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402411 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402442 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402484 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402506 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402535 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402564 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402588 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402643 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402679 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402718 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402744 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402775 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402797 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402823 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402842 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402867 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402888 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402906 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402931 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402951 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402976 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.402997 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403016 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403040 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403063 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403088 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403115 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403138 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403165 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403192 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403211 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403267 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403288 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403312 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403333 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403356 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403393 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403414 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403440 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403461 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403483 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403511 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403531 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403558 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403578 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403597 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403624 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403644 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403669 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403738 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403759 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403785 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403805 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403833 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403852 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403872 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403916 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403936 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403961 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.403983 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404003 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404029 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404050 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404073 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404099 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404119 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.404153 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.398638 4829 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a9b589fea7fe1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 07:16:49.383972833 +0000 UTC m=+0.723621298,LastTimestamp:2025-10-02 07:16:49.383972833 +0000 UTC m=+0.723621298,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.407834 4829 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.407911 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.407938 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.407989 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408014 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408042 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408064 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408085 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408111 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408132 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408154 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408182 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408204 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408279 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408300 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408321 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408348 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408367 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408394 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408416 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408453 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408481 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408502 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408527 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408546 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408570 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408598 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408631 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408657 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408677 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408697 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408722 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408742 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.408773 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.409086 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.409128 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.409192 4829 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.409220 4829 reconstruct.go:97] "Volume reconstruction finished" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.409270 4829 reconciler.go:26] "Reconciler: start to sync state" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.416658 4829 manager.go:324] Recovery completed Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.433722 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.435056 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.435104 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.435124 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.436558 4829 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.436577 4829 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.436603 4829 state_mem.go:36] "Initialized new in-memory state store" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.452724 4829 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.454376 4829 policy_none.go:49] "None policy: Start" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.455654 4829 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.455690 4829 state_mem.go:35] "Initializing new in-memory state store" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.459056 4829 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.459220 4829 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.459334 4829 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.459718 4829 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.460289 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.460366 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.487701 4829 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.522725 4829 manager.go:334] "Starting Device Plugin manager" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.522780 4829 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.522795 4829 server.go:79] "Starting device plugin registration server" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.523292 4829 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.523307 4829 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.524328 4829 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.524447 4829 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.524457 4829 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.535390 4829 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.560641 4829 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.560745 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.563656 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.563715 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.563733 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.563936 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.564282 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.564355 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565428 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565510 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565521 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565677 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.565893 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566003 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566481 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566504 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566515 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566607 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566789 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.566835 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567104 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567135 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567147 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567604 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567626 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567640 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567763 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567783 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.567844 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.568626 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.568651 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.568663 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.568816 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.568854 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.569195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.569218 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.569244 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.569588 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.569608 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.569619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.589613 4829 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="400ms" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611581 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611611 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611636 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611656 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611691 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611712 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611731 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611747 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611765 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611787 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611805 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611827 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611846 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611863 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.611883 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.623385 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.624208 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.624263 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.624283 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.624315 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.624811 4829 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713172 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713282 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713323 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713357 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713388 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713418 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713447 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713475 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713507 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713536 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713567 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713600 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713630 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713658 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713687 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713909 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.713984 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714012 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714373 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714415 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714011 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714485 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714532 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714574 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714783 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714797 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.714826 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.715021 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.715136 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.825287 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.827187 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.827350 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.827437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.827533 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.828023 4829 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.906738 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.917327 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.948187 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.952287 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f1ce4f90a38ec83015d3b4024364d33464e203a4ebdc25af5f2bfa5dcaab9ded WatchSource:0}: Error finding container f1ce4f90a38ec83015d3b4024364d33464e203a4ebdc25af5f2bfa5dcaab9ded: Status 404 returned error can't find the container with id f1ce4f90a38ec83015d3b4024364d33464e203a4ebdc25af5f2bfa5dcaab9ded Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.959515 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5dc6dbe980559b71da76de610cadf5ec1c902eb7f5a23ef01e81e9d3625aec82 WatchSource:0}: Error finding container 5dc6dbe980559b71da76de610cadf5ec1c902eb7f5a23ef01e81e9d3625aec82: Status 404 returned error can't find the container with id 5dc6dbe980559b71da76de610cadf5ec1c902eb7f5a23ef01e81e9d3625aec82 Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.976120 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: W1002 07:16:49.980030 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-18a45fc6f8ce6c8885c1563360299ed3dc136d510ce9db7e63f230396496a7b6 WatchSource:0}: Error finding container 18a45fc6f8ce6c8885c1563360299ed3dc136d510ce9db7e63f230396496a7b6: Status 404 returned error can't find the container with id 18a45fc6f8ce6c8885c1563360299ed3dc136d510ce9db7e63f230396496a7b6 Oct 02 07:16:49 crc kubenswrapper[4829]: I1002 07:16:49.984459 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:49 crc kubenswrapper[4829]: E1002 07:16:49.990703 4829 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="800ms" Oct 02 07:16:50 crc kubenswrapper[4829]: W1002 07:16:50.000732 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-2ced126c9ce0b6df9f5a08e3c7a2583ea05ac5cb735fb188faecc271327fd570 WatchSource:0}: Error finding container 2ced126c9ce0b6df9f5a08e3c7a2583ea05ac5cb735fb188faecc271327fd570: Status 404 returned error can't find the container with id 2ced126c9ce0b6df9f5a08e3c7a2583ea05ac5cb735fb188faecc271327fd570 Oct 02 07:16:50 crc kubenswrapper[4829]: W1002 07:16:50.002294 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-73fe2882b212d0ac3ebcc81857bb4bc3201a3dfec2d66f46ca9c319bc1f44fad WatchSource:0}: Error finding container 73fe2882b212d0ac3ebcc81857bb4bc3201a3dfec2d66f46ca9c319bc1f44fad: Status 404 returned error can't find the container with id 73fe2882b212d0ac3ebcc81857bb4bc3201a3dfec2d66f46ca9c319bc1f44fad Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.228582 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.230082 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.230120 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.230133 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.230157 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:50 crc kubenswrapper[4829]: E1002 07:16:50.230587 4829 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 02 07:16:50 crc kubenswrapper[4829]: W1002 07:16:50.332851 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:50 crc kubenswrapper[4829]: E1002 07:16:50.332977 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.385414 4829 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.466137 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"18a45fc6f8ce6c8885c1563360299ed3dc136d510ce9db7e63f230396496a7b6"} Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.467677 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5dc6dbe980559b71da76de610cadf5ec1c902eb7f5a23ef01e81e9d3625aec82"} Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.468708 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f1ce4f90a38ec83015d3b4024364d33464e203a4ebdc25af5f2bfa5dcaab9ded"} Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.469701 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"73fe2882b212d0ac3ebcc81857bb4bc3201a3dfec2d66f46ca9c319bc1f44fad"} Oct 02 07:16:50 crc kubenswrapper[4829]: I1002 07:16:50.470706 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2ced126c9ce0b6df9f5a08e3c7a2583ea05ac5cb735fb188faecc271327fd570"} Oct 02 07:16:50 crc kubenswrapper[4829]: W1002 07:16:50.525412 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:50 crc kubenswrapper[4829]: E1002 07:16:50.525502 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:50 crc kubenswrapper[4829]: E1002 07:16:50.791495 4829 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="1.6s" Oct 02 07:16:50 crc kubenswrapper[4829]: W1002 07:16:50.966484 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:50 crc kubenswrapper[4829]: E1002 07:16:50.966597 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:50 crc kubenswrapper[4829]: W1002 07:16:50.967996 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:50 crc kubenswrapper[4829]: E1002 07:16:50.968071 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.031100 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.033588 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.033648 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.033670 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.033710 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:51 crc kubenswrapper[4829]: E1002 07:16:51.034400 4829 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 02 07:16:51 crc kubenswrapper[4829]: E1002 07:16:51.139153 4829 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.200:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a9b589fea7fe1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 07:16:49.383972833 +0000 UTC m=+0.723621298,LastTimestamp:2025-10-02 07:16:49.383972833 +0000 UTC m=+0.723621298,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.385586 4829 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.476552 4829 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1" exitCode=0 Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.476879 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.476875 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.478841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.478900 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.478946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.482791 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.482869 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.482893 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.485593 4829 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74" exitCode=0 Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.485671 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.485784 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.487165 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.487204 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.487248 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.490212 4829 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653" exitCode=0 Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.490435 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.490972 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.491070 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.491619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.491655 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.491670 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.492360 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.492411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.492433 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.494236 4829 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="a11daceac317a87dfef15bdc352d73221a0793f6951cde667b9d92349f9e97cf" exitCode=0 Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.494277 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"a11daceac317a87dfef15bdc352d73221a0793f6951cde667b9d92349f9e97cf"} Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.494362 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.495544 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.495564 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:51 crc kubenswrapper[4829]: I1002 07:16:51.495571 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4829]: W1002 07:16:52.171331 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:52 crc kubenswrapper[4829]: E1002 07:16:52.171456 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.385410 4829 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:52 crc kubenswrapper[4829]: E1002 07:16:52.393312 4829 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.200:6443: connect: connection refused" interval="3.2s" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.499785 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.499830 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.499840 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.499853 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.503901 4829 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3" exitCode=0 Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.503968 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.504095 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.505017 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.505045 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.505057 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.513128 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f9691cc2ca2cc22e3f216ecc263a80476389bf663aa9a157e945a31f120b6474"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.513255 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.514183 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.514250 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.514267 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.523812 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.523873 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.523888 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.523907 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.524863 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.524882 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.524890 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.535059 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf"} Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.535258 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.536173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.536201 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.536215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.635345 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.637292 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.637355 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.637367 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:52 crc kubenswrapper[4829]: I1002 07:16:52.637385 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:52 crc kubenswrapper[4829]: E1002 07:16:52.637727 4829 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.200:6443: connect: connection refused" node="crc" Oct 02 07:16:52 crc kubenswrapper[4829]: W1002 07:16:52.838618 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.200:6443: connect: connection refused Oct 02 07:16:52 crc kubenswrapper[4829]: E1002 07:16:52.838725 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.200:6443: connect: connection refused" logger="UnhandledError" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.540773 4829 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa" exitCode=0 Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.540917 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa"} Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.541153 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.542218 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.542279 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.542310 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.546443 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.546485 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.546601 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.546435 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159"} Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.548018 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.548189 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.548690 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.548749 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.548770 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.549498 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.549523 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.549541 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.550413 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.550514 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.550555 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.552731 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.552787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.552808 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.799924 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:53 crc kubenswrapper[4829]: I1002 07:16:53.965592 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.499567 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.553030 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b"} Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.553077 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab"} Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.553093 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81"} Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.553055 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.553168 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.554732 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.554759 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.554815 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.554817 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.554849 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:54 crc kubenswrapper[4829]: I1002 07:16:54.554865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.563315 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0"} Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.563399 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4"} Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.563421 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.563473 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.565303 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.565368 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.565372 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.565400 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.565429 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.565449 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.838181 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.839970 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.840021 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.840040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:55 crc kubenswrapper[4829]: I1002 07:16:55.840072 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:16:56 crc kubenswrapper[4829]: I1002 07:16:56.566407 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:56 crc kubenswrapper[4829]: I1002 07:16:56.567814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:56 crc kubenswrapper[4829]: I1002 07:16:56.567879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:56 crc kubenswrapper[4829]: I1002 07:16:56.567899 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:57 crc kubenswrapper[4829]: I1002 07:16:57.423685 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:16:57 crc kubenswrapper[4829]: I1002 07:16:57.423927 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:57 crc kubenswrapper[4829]: I1002 07:16:57.429135 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:57 crc kubenswrapper[4829]: I1002 07:16:57.429259 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:57 crc kubenswrapper[4829]: I1002 07:16:57.429292 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.080606 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.080850 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.082663 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.082742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.082762 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.222293 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.515842 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.516110 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.518270 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.518322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.518340 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.524158 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:59 crc kubenswrapper[4829]: E1002 07:16:59.535569 4829 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.574495 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.574555 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.574591 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.576495 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.576572 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.576595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.576494 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.576699 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.576723 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:16:59 crc kubenswrapper[4829]: I1002 07:16:59.983726 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:17:00 crc kubenswrapper[4829]: I1002 07:17:00.411445 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:17:00 crc kubenswrapper[4829]: I1002 07:17:00.497452 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:17:00 crc kubenswrapper[4829]: I1002 07:17:00.577286 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:00 crc kubenswrapper[4829]: I1002 07:17:00.578786 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:00 crc kubenswrapper[4829]: I1002 07:17:00.578845 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:00 crc kubenswrapper[4829]: I1002 07:17:00.578865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:01 crc kubenswrapper[4829]: I1002 07:17:01.580136 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:01 crc kubenswrapper[4829]: I1002 07:17:01.581518 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:01 crc kubenswrapper[4829]: I1002 07:17:01.581567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:01 crc kubenswrapper[4829]: I1002 07:17:01.581586 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:02 crc kubenswrapper[4829]: I1002 07:17:02.984738 4829 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 07:17:02 crc kubenswrapper[4829]: I1002 07:17:02.984825 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.113163 4829 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:55148->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.113273 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:55148->192.168.126.11:17697: read: connection reset by peer" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.385348 4829 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.587784 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.589959 4829 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159" exitCode=255 Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.590020 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159"} Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.590137 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.590971 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.591017 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.591029 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.591645 4829 scope.go:117] "RemoveContainer" containerID="7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159" Oct 02 07:17:03 crc kubenswrapper[4829]: W1002 07:17:03.708064 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.708182 4829 trace.go:236] Trace[297939683]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:53.706) (total time: 10002ms): Oct 02 07:17:03 crc kubenswrapper[4829]: Trace[297939683]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (07:17:03.708) Oct 02 07:17:03 crc kubenswrapper[4829]: Trace[297939683]: [10.002114828s] [10.002114828s] END Oct 02 07:17:03 crc kubenswrapper[4829]: E1002 07:17:03.708213 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 07:17:03 crc kubenswrapper[4829]: W1002 07:17:03.994064 4829 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 07:17:03 crc kubenswrapper[4829]: I1002 07:17:03.994154 4829 trace.go:236] Trace[1749589439]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:53.992) (total time: 10001ms): Oct 02 07:17:03 crc kubenswrapper[4829]: Trace[1749589439]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:17:03.994) Oct 02 07:17:03 crc kubenswrapper[4829]: Trace[1749589439]: [10.00126259s] [10.00126259s] END Oct 02 07:17:03 crc kubenswrapper[4829]: E1002 07:17:03.994177 4829 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.339298 4829 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.339368 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.344792 4829 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.344827 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.595655 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.597900 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621"} Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.598075 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.599073 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.599133 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:04 crc kubenswrapper[4829]: I1002 07:17:04.599154 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.432876 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.433125 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.433278 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.435261 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.435324 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.435353 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.439324 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.605720 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.607010 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.607069 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.607091 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:07 crc kubenswrapper[4829]: I1002 07:17:07.810666 4829 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 07:17:08 crc kubenswrapper[4829]: I1002 07:17:08.608827 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:08 crc kubenswrapper[4829]: I1002 07:17:08.609979 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:08 crc kubenswrapper[4829]: I1002 07:17:08.610024 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:08 crc kubenswrapper[4829]: I1002 07:17:08.610032 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.256958 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.257220 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.258745 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.258830 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.258842 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.277463 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.339886 4829 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.343513 4829 trace.go:236] Trace[1313254332]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:56.626) (total time: 12717ms): Oct 02 07:17:09 crc kubenswrapper[4829]: Trace[1313254332]: ---"Objects listed" error: 12717ms (07:17:09.343) Oct 02 07:17:09 crc kubenswrapper[4829]: Trace[1313254332]: [12.71742515s] [12.71742515s] END Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.343571 4829 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.347152 4829 trace.go:236] Trace[657532891]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 07:16:58.477) (total time: 10870ms): Oct 02 07:17:09 crc kubenswrapper[4829]: Trace[657532891]: ---"Objects listed" error: 10870ms (07:17:09.347) Oct 02 07:17:09 crc kubenswrapper[4829]: Trace[657532891]: [10.870051411s] [10.870051411s] END Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.347198 4829 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.347369 4829 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.348216 4829 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.379094 4829 apiserver.go:52] "Watching apiserver" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.386221 4829 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.386578 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.387077 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.387144 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.387183 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.387583 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.387650 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.387736 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.387802 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.387828 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.388349 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.391740 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.391803 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.391842 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.391877 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.392034 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.393024 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.394859 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.395497 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.398814 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.427784 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.447912 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.447978 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.448020 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.448048 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.448080 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.449456 4829 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.459375 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.460130 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.465828 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.465870 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.465893 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.465984 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:09.965953777 +0000 UTC m=+21.305602222 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.479511 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.479574 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.479606 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.479699 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:09.979671201 +0000 UTC m=+21.319319646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.488075 4829 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.496287 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.507686 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.546073 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548453 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548528 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548575 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548614 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548659 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548701 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548736 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548768 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548839 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548882 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548896 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548923 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.548960 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549050 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549093 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549130 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549163 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549218 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549283 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549294 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549316 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549376 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549408 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549392 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549438 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549498 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549499 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549530 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549543 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549565 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549599 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549633 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549664 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549720 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549700 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549791 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549796 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549817 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549844 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549869 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549890 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549913 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549936 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549957 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.549980 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550006 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550015 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550041 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550064 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550088 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550109 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550130 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550153 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550179 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550200 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550242 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550264 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550287 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550308 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550329 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550381 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550404 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550427 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550453 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550476 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550499 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550523 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550546 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550569 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550622 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550646 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550668 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550690 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550722 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550746 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550771 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550793 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550814 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550836 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550863 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550886 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550912 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550933 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550955 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550980 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551029 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551052 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551076 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551099 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551121 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551179 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551205 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551247 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551271 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551293 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551319 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551343 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551366 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551389 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551411 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551502 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551528 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551550 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551573 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551596 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551619 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551642 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551664 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551686 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551709 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551755 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551781 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551804 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551825 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551848 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551873 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551897 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551922 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551945 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551968 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551991 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552015 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552036 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552058 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552081 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552104 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552127 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552154 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550122 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550200 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550421 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550452 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550558 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550720 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550791 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550856 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.550930 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551019 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551052 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551172 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551206 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551585 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551768 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551866 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.551955 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552164 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552567 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552177 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552712 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552735 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552760 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552782 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552803 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552825 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552849 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552872 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552895 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552918 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552927 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552942 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552965 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.552988 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553011 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553006 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553035 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553059 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553083 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553106 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553130 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553154 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553176 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553219 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553260 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553285 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553313 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553336 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553358 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553381 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553402 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553425 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553450 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553475 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553497 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553521 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553545 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553601 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553633 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553665 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553696 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553728 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553762 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553807 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553832 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553865 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553899 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553931 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553966 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554001 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554036 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554074 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554110 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554143 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554176 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554208 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554267 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554298 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554330 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554392 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554425 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554456 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554488 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554512 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554538 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554561 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554583 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554608 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554631 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554656 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554680 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554707 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.557512 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.557572 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.557599 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.560486 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553260 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.568664 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.553466 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554020 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554375 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554465 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554482 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554746 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554911 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.554996 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555046 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555140 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555177 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555282 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555430 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555438 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555574 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555582 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555708 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555744 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555811 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555840 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555899 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.555977 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.556031 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.556163 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.556163 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.556298 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.557521 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.557635 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:10.057615443 +0000 UTC m=+21.397263848 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.557987 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558091 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558182 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558295 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558417 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558509 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558590 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558655 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558759 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558762 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558764 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.558961 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.559133 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.559256 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.559430 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.559618 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.559942 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.560110 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.560276 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.560336 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.562535 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.564383 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.567333 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.567473 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.567699 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.567763 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.567911 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.568465 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.568636 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.568880 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.569010 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.569727 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.569762 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570045 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570108 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570258 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570512 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570574 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570755 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570799 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.570906 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.571148 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.571198 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.571416 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.571937 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.572029 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.572100 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.572368 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.572593 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.572784 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.572851 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.573363 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.573602 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.573808 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.574015 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.574512 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.575077 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.575150 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.575500 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.576289 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.576909 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.577155 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.576622 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.577625 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.590881 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.590940 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.591029 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.592643 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.592706 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597117 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597258 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597387 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597432 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597471 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597506 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597572 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.597636 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.599599 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.599613 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.599708 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:10.09968756 +0000 UTC m=+21.439335965 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.599969 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.602415 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.603294 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.603527 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.603734 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.603911 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.604507 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.604672 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.604970 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.605276 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.605522 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.605701 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.607958 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.608646 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.616088 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.616506 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.616607 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.616833 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.616883 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.617012 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.617477 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.618584 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.618917 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.625947 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.631105 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.631236 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: E1002 07:17:09.631379 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:10.131357042 +0000 UTC m=+21.471005447 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.631473 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.631509 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.631899 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.631963 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.630863 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.633037 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.633196 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.633741 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634386 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634833 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634855 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634868 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634886 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634899 4829 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634909 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634920 4829 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.634932 4829 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635021 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635064 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635077 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635091 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635132 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635149 4829 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635162 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635179 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635196 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635209 4829 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635253 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635274 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.631020 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635287 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635565 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635578 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635589 4829 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635600 4829 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635610 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635622 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635632 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635641 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635651 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635663 4829 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635673 4829 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635980 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.636246 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.636698 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.635683 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637144 4829 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637155 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637164 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637175 4829 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637605 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637623 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637636 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637651 4829 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637699 4829 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637708 4829 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637718 4829 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637740 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637751 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637761 4829 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637770 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637781 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637789 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637798 4829 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637809 4829 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637818 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637827 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637837 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637848 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637857 4829 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637865 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637874 4829 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637885 4829 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637894 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637904 4829 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637913 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637925 4829 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637936 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637944 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637955 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637970 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637978 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637986 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.637997 4829 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638006 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638016 4829 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638025 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638037 4829 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638046 4829 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638056 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638065 4829 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638077 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638086 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638095 4829 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638106 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638116 4829 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638125 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638135 4829 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638149 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638158 4829 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638167 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638177 4829 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638215 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638257 4829 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638267 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638294 4829 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638326 4829 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638340 4829 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638352 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638599 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638756 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638772 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.638690 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.639099 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.639561 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.640300 4829 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.640894 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.642248 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.647078 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.655493 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.655545 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.655819 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.656200 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.656484 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.661191 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.663512 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.663804 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.664268 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.664828 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.665512 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.665534 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.665694 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.665962 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.665969 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.666040 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.666923 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.667207 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.667234 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.668833 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.668963 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.669048 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.670614 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.670679 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.670874 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.670880 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.673345 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.673510 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.674247 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.676104 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.685041 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.685425 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.693319 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.693672 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.701311 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.709496 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.714164 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.717667 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:09 crc kubenswrapper[4829]: W1002 07:17:09.725903 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-1091df311114ef6ae71d9e8275d816e7de9cbe4361f95befbb7c496130282a5a WatchSource:0}: Error finding container 1091df311114ef6ae71d9e8275d816e7de9cbe4361f95befbb7c496130282a5a: Status 404 returned error can't find the container with id 1091df311114ef6ae71d9e8275d816e7de9cbe4361f95befbb7c496130282a5a Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.738321 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.738981 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739016 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739026 4829 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739034 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739042 4829 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739051 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739059 4829 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739067 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739075 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739083 4829 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739091 4829 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739099 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739107 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739115 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739123 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739133 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739142 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739151 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739159 4829 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739166 4829 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739174 4829 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739182 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739191 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739199 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739207 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739215 4829 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739242 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739253 4829 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739262 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739270 4829 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739278 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739286 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739294 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739301 4829 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739309 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739317 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739328 4829 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739336 4829 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739345 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739352 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739365 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739374 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739381 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739389 4829 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739397 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739405 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739412 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739420 4829 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739428 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739436 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739444 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739451 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739459 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739467 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739475 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739485 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739492 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739500 4829 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739508 4829 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739518 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739526 4829 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739534 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739542 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739550 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739559 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739567 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739575 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739583 4829 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739592 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739600 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739609 4829 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739617 4829 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739626 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739635 4829 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739644 4829 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739652 4829 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739660 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739669 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739677 4829 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739685 4829 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739693 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739701 4829 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739709 4829 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739716 4829 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739724 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739732 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739740 4829 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739747 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739756 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739765 4829 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739774 4829 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739782 4829 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739789 4829 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739797 4829 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739810 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739818 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739830 4829 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739838 4829 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739846 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739853 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.739892 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.750913 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.755531 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 07:17:09 crc kubenswrapper[4829]: W1002 07:17:09.764640 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-034b8c80b1e9739bf41ea0e7962b01cbc7478cc334a2619466d51419b01a8bc1 WatchSource:0}: Error finding container 034b8c80b1e9739bf41ea0e7962b01cbc7478cc334a2619466d51419b01a8bc1: Status 404 returned error can't find the container with id 034b8c80b1e9739bf41ea0e7962b01cbc7478cc334a2619466d51419b01a8bc1 Oct 02 07:17:09 crc kubenswrapper[4829]: I1002 07:17:09.994507 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.001779 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.007826 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.018099 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.027093 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.028081 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.042585 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.042660 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.042829 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.042856 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.042876 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.042933 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.042915968 +0000 UTC m=+22.382564383 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.043456 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.043495 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.043513 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.043588 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.043562627 +0000 UTC m=+22.383211252 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.054940 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.068533 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.078636 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.095500 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.110648 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.120854 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.133999 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.143291 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.143430 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.143468 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.143533 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.143498606 +0000 UTC m=+22.483147031 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.143554 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.143625 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.143602239 +0000 UTC m=+22.483250644 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.143675 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.143719 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:11.143710012 +0000 UTC m=+22.483358407 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.145996 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.161317 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.170878 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.179882 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.198001 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.645809 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"034b8c80b1e9739bf41ea0e7962b01cbc7478cc334a2619466d51419b01a8bc1"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.648549 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.648603 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.648618 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cec5cca91f3297e2b350b346d807319e409cb629dd55e9a8574861bc70ffea7a"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.650119 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.650202 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1091df311114ef6ae71d9e8275d816e7de9cbe4361f95befbb7c496130282a5a"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.651698 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.652173 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.654336 4829 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621" exitCode=255 Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.654411 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621"} Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.654472 4829 scope.go:117] "RemoveContainer" containerID="7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.669597 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.669932 4829 scope.go:117] "RemoveContainer" containerID="ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621" Oct 02 07:17:10 crc kubenswrapper[4829]: E1002 07:17:10.670107 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.674114 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.691561 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.703818 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.716685 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.727693 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.747315 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.766347 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.783176 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.795930 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.809272 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.825618 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:03Z\\\",\\\"message\\\":\\\"W1002 07:16:52.637187 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 07:16:52.637782 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759389412 cert, and key in /tmp/serving-cert-3258558389/serving-signer.crt, /tmp/serving-cert-3258558389/serving-signer.key\\\\nI1002 07:16:52.823843 1 observer_polling.go:159] Starting file observer\\\\nW1002 07:16:52.827280 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 07:16:52.827645 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:16:52.829464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3258558389/tls.crt::/tmp/serving-cert-3258558389/tls.key\\\\\\\"\\\\nF1002 07:17:03.107582 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.836168 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.849089 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.866040 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.884322 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.897328 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:10 crc kubenswrapper[4829]: I1002 07:17:10.918705 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:10Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.051481 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.051596 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.051799 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.051829 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.051849 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.051942 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:13.051919746 +0000 UTC m=+24.391568181 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.052500 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.052531 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.052549 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.052594 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:13.052578926 +0000 UTC m=+24.392227361 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.152707 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.152826 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.152870 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.152983 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.153053 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:13.153031311 +0000 UTC m=+24.492679746 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.153584 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:13.153565166 +0000 UTC m=+24.493213611 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.153692 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.153737 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:13.153724211 +0000 UTC m=+24.493372656 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.158777 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2lvjk"] Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.159894 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.162246 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-xrncc"] Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.162586 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.162689 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.162625 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.162633 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.164091 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-vshhb"] Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.164338 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.164803 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-spzjt"] Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.165021 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.165577 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.168170 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.168298 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.168549 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.168653 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.168757 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.168595 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169094 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169127 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169338 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-8rvq4"] Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169496 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169627 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169731 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.169916 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.170878 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.172435 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.176181 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.176246 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.176285 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.176196 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.176214 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.199176 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:03Z\\\",\\\"message\\\":\\\"W1002 07:16:52.637187 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 07:16:52.637782 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759389412 cert, and key in /tmp/serving-cert-3258558389/serving-signer.crt, /tmp/serving-cert-3258558389/serving-signer.key\\\\nI1002 07:16:52.823843 1 observer_polling.go:159] Starting file observer\\\\nW1002 07:16:52.827280 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 07:16:52.827645 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:16:52.829464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3258558389/tls.crt::/tmp/serving-cert-3258558389/tls.key\\\\\\\"\\\\nF1002 07:17:03.107582 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.221928 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.236761 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.252169 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.253398 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-os-release\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.253575 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9jlp\" (UniqueName: \"kubernetes.io/projected/60aa0a81-354e-46f1-ab0c-b1eb386974a6-kube-api-access-l9jlp\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.253780 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-node-log\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.253926 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-env-overrides\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254080 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/87f69905-b2b5-4012-9b1d-77e8d22b2cb3-hosts-file\") pod \"node-resolver-vshhb\" (UID: \"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\") " pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254189 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-hostroot\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254596 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ca00dadf-1664-466b-830c-e172857db47a-cni-binary-copy\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254753 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/60aa0a81-354e-46f1-ab0c-b1eb386974a6-proxy-tls\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254792 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60aa0a81-354e-46f1-ab0c-b1eb386974a6-mcd-auth-proxy-config\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254877 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-systemd-units\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254924 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-netns\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.254966 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-etc-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258022 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-system-cni-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258066 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258100 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4smrk\" (UniqueName: \"kubernetes.io/projected/00c0190d-cefe-408f-88c0-fa849e3811d3-kube-api-access-4smrk\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258138 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-script-lib\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258170 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-socket-dir-parent\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258202 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-kubelet\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258286 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ca00dadf-1664-466b-830c-e172857db47a-multus-daemon-config\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258325 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cni-binary-copy\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258358 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258384 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/60aa0a81-354e-46f1-ab0c-b1eb386974a6-rootfs\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258424 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-multus-certs\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258456 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-kubelet\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258482 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-var-lib-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258512 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-netd\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258537 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qt7q\" (UniqueName: \"kubernetes.io/projected/87f69905-b2b5-4012-9b1d-77e8d22b2cb3-kube-api-access-7qt7q\") pod \"node-resolver-vshhb\" (UID: \"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\") " pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258567 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-os-release\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258596 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-system-cni-dir\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258622 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-log-socket\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258644 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258675 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-cni-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258706 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-cnibin\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258737 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-conf-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.258984 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-cni-bin\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259055 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-systemd\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259092 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-netns\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259136 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whzpm\" (UniqueName: \"kubernetes.io/projected/ca00dadf-1664-466b-830c-e172857db47a-kube-api-access-whzpm\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259175 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-bin\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259362 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-config\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259605 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00c0190d-cefe-408f-88c0-fa849e3811d3-ovn-node-metrics-cert\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.259983 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-etc-kubernetes\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260176 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260343 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260468 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-slash\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260609 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-k8s-cni-cncf-io\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260665 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cnibin\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260717 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-cni-multus\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsfzw\" (UniqueName: \"kubernetes.io/projected/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-kube-api-access-zsfzw\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.260822 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-ovn\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.270605 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.289108 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.311518 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.328635 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.342523 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.354026 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362173 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-os-release\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362256 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9jlp\" (UniqueName: \"kubernetes.io/projected/60aa0a81-354e-46f1-ab0c-b1eb386974a6-kube-api-access-l9jlp\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362304 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-node-log\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362337 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-env-overrides\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362370 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/87f69905-b2b5-4012-9b1d-77e8d22b2cb3-hosts-file\") pod \"node-resolver-vshhb\" (UID: \"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\") " pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362399 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-hostroot\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362428 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ca00dadf-1664-466b-830c-e172857db47a-cni-binary-copy\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362457 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/60aa0a81-354e-46f1-ab0c-b1eb386974a6-proxy-tls\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362486 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60aa0a81-354e-46f1-ab0c-b1eb386974a6-mcd-auth-proxy-config\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362541 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-systemd-units\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362568 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-netns\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362599 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-etc-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362633 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-hostroot\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362642 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-system-cni-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362706 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362722 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-system-cni-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362725 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4smrk\" (UniqueName: \"kubernetes.io/projected/00c0190d-cefe-408f-88c0-fa849e3811d3-kube-api-access-4smrk\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362783 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-script-lib\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362820 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-socket-dir-parent\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362853 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-kubelet\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362883 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ca00dadf-1664-466b-830c-e172857db47a-multus-daemon-config\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362920 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cni-binary-copy\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362950 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362980 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/60aa0a81-354e-46f1-ab0c-b1eb386974a6-rootfs\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363012 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-multus-certs\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363042 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-kubelet\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363077 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-var-lib-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363087 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-node-log\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363108 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-netd\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363142 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qt7q\" (UniqueName: \"kubernetes.io/projected/87f69905-b2b5-4012-9b1d-77e8d22b2cb3-kube-api-access-7qt7q\") pod \"node-resolver-vshhb\" (UID: \"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\") " pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363173 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-os-release\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363202 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-system-cni-dir\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363255 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-log-socket\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363286 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363321 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-cni-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363350 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-cnibin\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363380 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-conf-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363438 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-cni-bin\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363469 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-systemd\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363499 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-netns\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363527 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whzpm\" (UniqueName: \"kubernetes.io/projected/ca00dadf-1664-466b-830c-e172857db47a-kube-api-access-whzpm\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363558 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-bin\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363589 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-config\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363609 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-env-overrides\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363618 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00c0190d-cefe-408f-88c0-fa849e3811d3-ovn-node-metrics-cert\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363665 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-etc-kubernetes\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363674 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/87f69905-b2b5-4012-9b1d-77e8d22b2cb3-hosts-file\") pod \"node-resolver-vshhb\" (UID: \"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\") " pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363698 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363708 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-ovn-kubernetes\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363729 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363740 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-netd\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363774 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-slash\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363806 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-k8s-cni-cncf-io\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363836 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cnibin\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363865 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-cni-multus\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363895 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsfzw\" (UniqueName: \"kubernetes.io/projected/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-kube-api-access-zsfzw\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.363925 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-ovn\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364030 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-ovn\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364218 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ca00dadf-1664-466b-830c-e172857db47a-cni-binary-copy\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364308 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-kubelet\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364473 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-os-release\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364543 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-system-cni-dir\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364593 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-log-socket\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364639 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364485 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-socket-dir-parent\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364785 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ca00dadf-1664-466b-830c-e172857db47a-multus-daemon-config\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.364855 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-cnibin\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.362596 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-os-release\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365312 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-conf-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365361 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-cni-bin\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365390 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-systemd\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365397 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cni-binary-copy\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365440 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-var-lib-cni-multus\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365417 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-netns\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365630 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-bin\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365773 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-netns\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.365952 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-multus-certs\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366030 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-systemd-units\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366065 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-etc-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366179 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-config\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366289 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/60aa0a81-354e-46f1-ab0c-b1eb386974a6-rootfs\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366375 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-kubelet\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366448 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-var-lib-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366520 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-openvswitch\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366583 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-etc-kubernetes\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366656 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-host-run-k8s-cni-cncf-io\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366728 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cnibin\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.366721 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-slash\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.367208 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-script-lib\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.367218 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.367251 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/60aa0a81-354e-46f1-ab0c-b1eb386974a6-mcd-auth-proxy-config\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.367434 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ca00dadf-1664-466b-830c-e172857db47a-multus-cni-dir\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.367435 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.373275 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.373856 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/60aa0a81-354e-46f1-ab0c-b1eb386974a6-proxy-tls\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.378543 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00c0190d-cefe-408f-88c0-fa849e3811d3-ovn-node-metrics-cert\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.384953 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4smrk\" (UniqueName: \"kubernetes.io/projected/00c0190d-cefe-408f-88c0-fa849e3811d3-kube-api-access-4smrk\") pod \"ovnkube-node-2lvjk\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.391035 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whzpm\" (UniqueName: \"kubernetes.io/projected/ca00dadf-1664-466b-830c-e172857db47a-kube-api-access-whzpm\") pod \"multus-spzjt\" (UID: \"ca00dadf-1664-466b-830c-e172857db47a\") " pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.396738 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.399410 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsfzw\" (UniqueName: \"kubernetes.io/projected/8880512f-81d6-4b87-82b7-4ce1a2a13f9a-kube-api-access-zsfzw\") pod \"multus-additional-cni-plugins-xrncc\" (UID: \"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\") " pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.400194 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qt7q\" (UniqueName: \"kubernetes.io/projected/87f69905-b2b5-4012-9b1d-77e8d22b2cb3-kube-api-access-7qt7q\") pod \"node-resolver-vshhb\" (UID: \"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\") " pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.401114 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9jlp\" (UniqueName: \"kubernetes.io/projected/60aa0a81-354e-46f1-ab0c-b1eb386974a6-kube-api-access-l9jlp\") pod \"machine-config-daemon-8rvq4\" (UID: \"60aa0a81-354e-46f1-ab0c-b1eb386974a6\") " pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.414452 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.425922 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.442983 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.458764 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.462464 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.462542 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.462675 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.462479 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.462833 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.462958 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.467150 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.467845 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.469369 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.470164 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.471592 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.472289 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.472992 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.474163 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.474842 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.475009 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.476280 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.476951 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.478850 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.479578 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.481376 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.482021 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.483121 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.483748 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.484157 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.485123 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.485506 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-xrncc" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.485599 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.485694 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.486119 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.487591 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.488032 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.489060 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.489500 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.491400 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vshhb" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.494079 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.495046 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.495537 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.497356 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-spzjt" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.499359 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.500339 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.503518 4829 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.503700 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.505562 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7d37f44d4abdf069737d20ab19989066c18849b0bfc0d9042190d81f12618159\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:03Z\\\",\\\"message\\\":\\\"W1002 07:16:52.637187 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1002 07:16:52.637782 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759389412 cert, and key in /tmp/serving-cert-3258558389/serving-signer.crt, /tmp/serving-cert-3258558389/serving-signer.key\\\\nI1002 07:16:52.823843 1 observer_polling.go:159] Starting file observer\\\\nW1002 07:16:52.827280 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1002 07:16:52.827645 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:16:52.829464 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3258558389/tls.crt::/tmp/serving-cert-3258558389/tls.key\\\\\\\"\\\\nF1002 07:17:03.107582 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.505866 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.517328 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.519373 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.521474 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: W1002 07:17:11.522408 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87f69905_b2b5_4012_9b1d_77e8d22b2cb3.slice/crio-9498e33ad5be4ffc284d4e8a3e360aa29bb4bd05954b6b69232acbddf97ffd9a WatchSource:0}: Error finding container 9498e33ad5be4ffc284d4e8a3e360aa29bb4bd05954b6b69232acbddf97ffd9a: Status 404 returned error can't find the container with id 9498e33ad5be4ffc284d4e8a3e360aa29bb4bd05954b6b69232acbddf97ffd9a Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.525429 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.527020 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.529811 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.531799 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.533775 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.534645 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.535020 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.536535 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.537682 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.539566 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.540460 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.541390 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.542776 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.545289 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.546111 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.546957 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.548641 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.549506 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.551149 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.552018 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 07:17:11 crc kubenswrapper[4829]: W1002 07:17:11.555815 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca00dadf_1664_466b_830c_e172857db47a.slice/crio-fc6a31ab22a61dd7115c4a7fd465bcfc00a210074d59a423c4f1da305d79f3f5 WatchSource:0}: Error finding container fc6a31ab22a61dd7115c4a7fd465bcfc00a210074d59a423c4f1da305d79f3f5: Status 404 returned error can't find the container with id fc6a31ab22a61dd7115c4a7fd465bcfc00a210074d59a423c4f1da305d79f3f5 Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.566565 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.612696 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.633352 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.658090 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.668108 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerStarted","Data":"94ca1575fac3fe796bf7b340e3f5e816bb2c69b1b82754bcbeac6398f93b1f0d"} Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.681362 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.685421 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.687906 4829 scope.go:117] "RemoveContainer" containerID="ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621" Oct 02 07:17:11 crc kubenswrapper[4829]: E1002 07:17:11.688146 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.689136 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vshhb" event={"ID":"87f69905-b2b5-4012-9b1d-77e8d22b2cb3","Type":"ContainerStarted","Data":"9498e33ad5be4ffc284d4e8a3e360aa29bb4bd05954b6b69232acbddf97ffd9a"} Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.693000 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"0ec09b6c64e5ce70c10c47472e451e836c791d55f25622d3490d257010b85834"} Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.697137 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"4d4e05b1a21d05512846aa2ef62bcb5860c0059a512017eeab37b7cc25c4739e"} Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.698543 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerStarted","Data":"fc6a31ab22a61dd7115c4a7fd465bcfc00a210074d59a423c4f1da305d79f3f5"} Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.704606 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.720077 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.730173 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.739676 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.750157 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.770873 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.783605 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.795285 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.805034 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.815063 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.833419 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.848825 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.860658 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:11 crc kubenswrapper[4829]: I1002 07:17:11.873362 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:11Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.081323 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.702331 4829 generic.go:334] "Generic (PLEG): container finished" podID="8880512f-81d6-4b87-82b7-4ce1a2a13f9a" containerID="3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310" exitCode=0 Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.702388 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerDied","Data":"3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.704752 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vshhb" event={"ID":"87f69905-b2b5-4012-9b1d-77e8d22b2cb3","Type":"ContainerStarted","Data":"1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.706163 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800" exitCode=0 Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.706298 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.716313 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.726444 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.726495 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.731197 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.739143 4829 scope.go:117] "RemoveContainer" containerID="ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621" Oct 02 07:17:12 crc kubenswrapper[4829]: E1002 07:17:12.739391 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.739532 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerStarted","Data":"983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428"} Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.767726 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.782416 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.792394 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.802292 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.825781 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.861532 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.894520 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.904188 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.913610 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.928266 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.939588 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.950936 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.966149 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.977733 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:12 crc kubenswrapper[4829]: I1002 07:17:12.987591 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:12Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.005339 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.018701 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.032371 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.046842 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.059252 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.073076 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.081468 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.081511 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081622 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081640 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081651 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081687 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:17.08167513 +0000 UTC m=+28.421323535 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081943 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081958 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081966 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.081988 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:17.081980639 +0000 UTC m=+28.421629044 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.088561 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.102723 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.115897 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.130551 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.152333 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.174215 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.182415 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.182530 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:17.182511256 +0000 UTC m=+28.522159661 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.182637 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.182761 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.182797 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:17.182787384 +0000 UTC m=+28.522435789 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.182873 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.182932 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:17.182917798 +0000 UTC m=+28.522566203 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.182673 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.460726 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.461173 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.461292 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.461374 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.461531 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:13 crc kubenswrapper[4829]: E1002 07:17:13.461705 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.746353 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b"} Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.746825 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c"} Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.746854 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d"} Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.746886 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844"} Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.746911 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0"} Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.748421 4829 generic.go:334] "Generic (PLEG): container finished" podID="8880512f-81d6-4b87-82b7-4ce1a2a13f9a" containerID="27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817" exitCode=0 Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.748670 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerDied","Data":"27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817"} Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.764280 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.787310 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.801686 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.818544 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.829258 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.845506 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.861157 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.874753 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.891750 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.909905 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.942282 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.973337 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:13 crc kubenswrapper[4829]: I1002 07:17:13.989564 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:13Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.007051 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.242658 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-tbzrs"] Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.243193 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.246095 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.247121 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.247504 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.248533 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.262845 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.278857 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.314609 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.328869 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.351547 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.370189 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.389433 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.396452 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9a0b933-69ad-4084-a133-df1872f73c29-host\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.396560 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f9a0b933-69ad-4084-a133-df1872f73c29-serviceca\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.396709 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k7pp\" (UniqueName: \"kubernetes.io/projected/f9a0b933-69ad-4084-a133-df1872f73c29-kube-api-access-4k7pp\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.411928 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.428984 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.448013 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.462306 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.475760 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.498346 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f9a0b933-69ad-4084-a133-df1872f73c29-serviceca\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.498499 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k7pp\" (UniqueName: \"kubernetes.io/projected/f9a0b933-69ad-4084-a133-df1872f73c29-kube-api-access-4k7pp\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.498578 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9a0b933-69ad-4084-a133-df1872f73c29-host\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.498677 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f9a0b933-69ad-4084-a133-df1872f73c29-host\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.501276 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f9a0b933-69ad-4084-a133-df1872f73c29-serviceca\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.510351 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.526392 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k7pp\" (UniqueName: \"kubernetes.io/projected/f9a0b933-69ad-4084-a133-df1872f73c29-kube-api-access-4k7pp\") pod \"node-ca-tbzrs\" (UID: \"f9a0b933-69ad-4084-a133-df1872f73c29\") " pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.534879 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.558804 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.562945 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tbzrs" Oct 02 07:17:14 crc kubenswrapper[4829]: W1002 07:17:14.589437 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9a0b933_69ad_4084_a133_df1872f73c29.slice/crio-9999bbdb76a98d7d3fcc7adcefc91c73722f51d36497423e37a4e11b3b2266e7 WatchSource:0}: Error finding container 9999bbdb76a98d7d3fcc7adcefc91c73722f51d36497423e37a4e11b3b2266e7: Status 404 returned error can't find the container with id 9999bbdb76a98d7d3fcc7adcefc91c73722f51d36497423e37a4e11b3b2266e7 Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.761488 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tbzrs" event={"ID":"f9a0b933-69ad-4084-a133-df1872f73c29","Type":"ContainerStarted","Data":"9999bbdb76a98d7d3fcc7adcefc91c73722f51d36497423e37a4e11b3b2266e7"} Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.770726 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95"} Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.774137 4829 generic.go:334] "Generic (PLEG): container finished" podID="8880512f-81d6-4b87-82b7-4ce1a2a13f9a" containerID="e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20" exitCode=0 Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.774191 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerDied","Data":"e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20"} Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.793377 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.811317 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.829382 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.847607 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.867894 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.899673 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.914896 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.935977 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.955072 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.972770 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:14 crc kubenswrapper[4829]: I1002 07:17:14.996375 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:14Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.010794 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.028015 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.042565 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.060212 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.460496 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.460607 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.460699 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.460723 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.460795 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.460944 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.748356 4829 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.755921 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.755985 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.756003 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.756132 4829 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.766168 4829 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.766562 4829 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.768685 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.768754 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.768772 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.768799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.768816 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.781808 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tbzrs" event={"ID":"f9a0b933-69ad-4084-a133-df1872f73c29","Type":"ContainerStarted","Data":"6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.787763 4829 generic.go:334] "Generic (PLEG): container finished" podID="8880512f-81d6-4b87-82b7-4ce1a2a13f9a" containerID="072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20" exitCode=0 Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.787825 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerDied","Data":"072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20"} Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.795785 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.800116 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.800156 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.800173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.800195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.800212 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.807135 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.826057 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.826907 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.833860 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.834077 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.834211 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.834441 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.834605 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.851359 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.854862 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.860261 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.860331 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.860350 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.860374 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.860393 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.866689 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.881668 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.886532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.886555 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.886564 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.886577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.886589 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.893577 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.903613 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: E1002 07:17:15.903810 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.905164 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.905188 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.905198 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.905214 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.905240 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:15Z","lastTransitionTime":"2025-10-02T07:17:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.914327 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.925661 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.934793 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.955685 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.968678 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.980036 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:15 crc kubenswrapper[4829]: I1002 07:17:15.997349 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:15Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.007410 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.007467 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.007485 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.007512 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.007531 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.010556 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.029436 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.045212 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.059056 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.073206 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.086490 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.100285 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.109977 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.110030 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.110054 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.110087 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.110113 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.114322 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.145770 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.162792 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.180370 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.193634 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.206426 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.213302 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.213345 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.213354 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.213369 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.213378 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.236314 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.257290 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.280250 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.292894 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.309291 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.315974 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.316026 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.316054 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.316080 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.316097 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.418422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.418797 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.418809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.418825 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.418835 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.532101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.532152 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.532169 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.532192 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.532209 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.636272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.636359 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.636385 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.636419 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.636444 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.739212 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.739291 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.739309 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.739332 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.739351 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.797771 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.802945 4829 generic.go:334] "Generic (PLEG): container finished" podID="8880512f-81d6-4b87-82b7-4ce1a2a13f9a" containerID="eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52" exitCode=0 Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.802980 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerDied","Data":"eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.830982 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.854399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.854506 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.854532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.854607 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.854625 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.854659 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.885194 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.905207 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.934301 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.952140 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.957983 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.959087 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.959344 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.959438 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.959460 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:16Z","lastTransitionTime":"2025-10-02T07:17:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.977052 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:16 crc kubenswrapper[4829]: I1002 07:17:16.991244 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:16Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.010261 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.026587 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.041609 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.060363 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.063290 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.063324 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.063334 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.063346 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.063358 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.094715 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.116674 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.137249 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.151655 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.151715 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151862 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151892 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151890 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151905 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151918 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151935 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151964 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:25.151947854 +0000 UTC m=+36.491596259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.151992 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:25.151973915 +0000 UTC m=+36.491622380 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.168927 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.168996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.169020 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.169051 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.169073 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.253094 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.253213 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.253293 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.253375 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.253428 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:25.253411629 +0000 UTC m=+36.593060044 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.253812 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:25.25380041 +0000 UTC m=+36.593448825 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.253959 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.254159 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:25.25412791 +0000 UTC m=+36.593776345 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.271937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.271997 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.272016 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.272048 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.272073 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.375865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.375934 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.375952 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.375978 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.376001 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.460050 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.460132 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.460179 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.460282 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.460448 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:17 crc kubenswrapper[4829]: E1002 07:17:17.461508 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.479033 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.479079 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.479094 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.479112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.479129 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.581168 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.581247 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.581267 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.581291 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.581309 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.683442 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.683518 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.683538 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.683563 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.683579 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.786186 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.786290 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.786311 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.786347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.786366 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.809763 4829 generic.go:334] "Generic (PLEG): container finished" podID="8880512f-81d6-4b87-82b7-4ce1a2a13f9a" containerID="ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3" exitCode=0 Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.809823 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerDied","Data":"ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.828458 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.850261 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.868730 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.880383 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.893262 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.893350 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.893377 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.893413 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.893437 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:17Z","lastTransitionTime":"2025-10-02T07:17:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.894160 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.907394 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.925115 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.942054 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.959278 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.975794 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:17 crc kubenswrapper[4829]: I1002 07:17:17.993499 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.004409 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.004455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.004490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.004510 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.004522 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.008160 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.032153 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.050393 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.066847 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.107786 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.107873 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.107897 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.107933 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.107957 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.209973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.210022 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.210038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.210060 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.210075 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.313915 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.313962 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.313973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.313992 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.314028 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.418117 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.418178 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.418197 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.418261 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.418288 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.521501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.521572 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.521599 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.521800 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.521823 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.625564 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.625632 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.625649 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.625673 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.625690 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.728323 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.728383 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.728402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.728430 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.728461 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.820056 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.820502 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.826876 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" event={"ID":"8880512f-81d6-4b87-82b7-4ce1a2a13f9a","Type":"ContainerStarted","Data":"a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.831785 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.831841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.831857 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.831881 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.831899 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.845622 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.866289 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.886128 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.896461 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.903003 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.922823 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.934751 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.934814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.934834 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.934861 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.934881 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:18Z","lastTransitionTime":"2025-10-02T07:17:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.956350 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.973310 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:18 crc kubenswrapper[4829]: I1002 07:17:18.995830 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:18Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.012071 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.032332 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.037437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.037464 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.037476 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.037507 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.037520 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.060311 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.075590 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.095650 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.112719 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.135957 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.141553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.141600 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.141614 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.141631 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.141642 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.169247 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.187024 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.207780 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.227058 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.244641 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.244701 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.244722 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.244754 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.244778 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.250265 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.284142 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.301892 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.322919 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.342045 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.347579 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.347656 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.347682 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.347713 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.347737 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.367198 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.393086 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.418509 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.441949 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.451197 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.451275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.451293 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.451318 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.451335 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.459844 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.459976 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.459843 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:19 crc kubenswrapper[4829]: E1002 07:17:19.460062 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:19 crc kubenswrapper[4829]: E1002 07:17:19.460164 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:19 crc kubenswrapper[4829]: E1002 07:17:19.460311 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.460457 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.478040 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.505580 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.525296 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.548766 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.553408 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.553456 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.553473 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.553498 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.553515 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.570187 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.584265 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.598717 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.615627 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.633434 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.656524 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.658802 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.658868 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.658891 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.658921 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.658946 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.677321 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.700591 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.715742 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.740516 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.754514 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.762139 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.762195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.762212 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.762266 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.762288 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.768917 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.830149 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.831102 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.866211 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.866363 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.866731 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.866756 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.866780 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.866798 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.882764 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.896761 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.916554 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.934740 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.952355 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.968977 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.969809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.969883 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.969906 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.969932 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:19 crc kubenswrapper[4829]: I1002 07:17:19.969950 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:19Z","lastTransitionTime":"2025-10-02T07:17:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.000551 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.019366 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.034900 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.053078 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.072330 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.072416 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.072439 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.072471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.072495 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.072888 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.102542 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.126555 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.146837 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.161455 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:20Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.175445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.175511 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.175532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.175557 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.175576 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.278759 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.279101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.279257 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.279375 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.279492 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.382510 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.382577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.382595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.382620 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.382639 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.486322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.486382 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.486403 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.486430 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.486453 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.589652 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.589695 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.589707 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.589725 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.589745 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.691654 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.691696 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.691708 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.691727 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.691738 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.797768 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.797809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.797819 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.797835 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.797847 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.833374 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.901540 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.901595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.901607 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.901641 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:20 crc kubenswrapper[4829]: I1002 07:17:20.901653 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:20Z","lastTransitionTime":"2025-10-02T07:17:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.005601 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.005905 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.006101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.006342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.006573 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.109445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.109502 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.109521 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.109544 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.109561 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.213020 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.213080 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.213097 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.213119 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.213137 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.315912 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.315974 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.315993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.316021 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.316041 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.419279 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.419338 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.419359 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.419383 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.419401 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.460663 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.460663 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:21 crc kubenswrapper[4829]: E1002 07:17:21.461075 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.460714 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:21 crc kubenswrapper[4829]: E1002 07:17:21.461196 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:21 crc kubenswrapper[4829]: E1002 07:17:21.461098 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.522294 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.522341 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.522359 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.522386 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.522404 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.625696 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.625772 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.625789 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.625816 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.625835 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.728944 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.728996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.729014 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.729040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.729056 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.831756 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.831803 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.831818 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.831841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.831857 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.837504 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/0.log" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.840323 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760" exitCode=1 Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.840357 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.841066 4829 scope.go:117] "RemoveContainer" containerID="8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.866294 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.889277 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:21Z\\\",\\\"message\\\":\\\"minpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290569 6153 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290824 6153 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:21.290916 6153 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:17:21.291269 6153 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:21.291293 6153 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:21.291314 6153 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:21.291315 6153 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:21.291323 6153 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:21.291360 6153 factory.go:656] Stopping watch factory\\\\nI1002 07:17:21.291378 6153 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:21.291381 6153 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:21.291389 6153 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.901589 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.919333 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.934853 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.934916 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.934935 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.934961 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.934979 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:21Z","lastTransitionTime":"2025-10-02T07:17:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.937016 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.951420 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.969551 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:21 crc kubenswrapper[4829]: I1002 07:17:21.985547 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:21Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.005050 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.020963 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.038019 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.038458 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.038489 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.038525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.038541 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.038901 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.057623 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.081013 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.097990 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.132525 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.142147 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.142206 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.142248 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.142322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.142353 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.244905 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.244946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.244957 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.244973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.244983 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.349049 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.349114 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.349133 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.349159 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.349182 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.452279 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.452382 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.452406 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.452437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.452459 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.556358 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.556418 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.556436 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.556461 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.556482 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.659144 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.659201 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.659213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.659248 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.659264 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.762982 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.763041 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.763059 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.763087 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.763129 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.846877 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/0.log" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.851130 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.851336 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.866127 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.866208 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.866264 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.866293 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.866314 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.878390 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.904208 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.919997 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.932738 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.943891 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.964333 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.969315 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.969357 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.969369 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.969386 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.969403 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:22Z","lastTransitionTime":"2025-10-02T07:17:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.981328 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:22 crc kubenswrapper[4829]: I1002 07:17:22.994437 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:22Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.009207 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.024077 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.054919 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:21Z\\\",\\\"message\\\":\\\"minpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290569 6153 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290824 6153 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:21.290916 6153 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:17:21.291269 6153 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:21.291293 6153 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:21.291314 6153 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:21.291315 6153 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:21.291323 6153 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:21.291360 6153 factory.go:656] Stopping watch factory\\\\nI1002 07:17:21.291378 6153 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:21.291381 6153 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:21.291389 6153 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.071725 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.071759 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.071769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.071784 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.071794 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.073509 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.087755 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.103037 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.118255 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.174217 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.174269 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.174281 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.174297 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.174307 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.277393 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.277470 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.277490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.277515 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.277533 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.380262 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.380330 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.380347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.380372 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.380396 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.460657 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.460709 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.460747 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:23 crc kubenswrapper[4829]: E1002 07:17:23.460862 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:23 crc kubenswrapper[4829]: E1002 07:17:23.460996 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:23 crc kubenswrapper[4829]: E1002 07:17:23.461107 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.483153 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.483266 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.483293 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.483322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.483343 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.586504 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.586559 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.586578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.586602 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.586619 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.690050 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.690110 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.690134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.690169 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.690192 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.793272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.793323 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.793339 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.793363 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.793382 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.838202 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc"] Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.838829 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.842186 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.842473 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.858797 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.860784 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/1.log" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.861883 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/0.log" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.866990 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f" exitCode=1 Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.867043 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.867131 4829 scope.go:117] "RemoveContainer" containerID="8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.868105 4829 scope.go:117] "RemoveContainer" containerID="1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f" Oct 02 07:17:23 crc kubenswrapper[4829]: E1002 07:17:23.868460 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.881731 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.895986 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.896022 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.896033 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.896053 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.896066 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.902370 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.910288 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.921176 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c45bd15-86f4-4d9c-8dfc-04274acd600f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.921269 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c45bd15-86f4-4d9c-8dfc-04274acd600f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.921296 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.921499 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c45bd15-86f4-4d9c-8dfc-04274acd600f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.921703 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24gn2\" (UniqueName: \"kubernetes.io/projected/8c45bd15-86f4-4d9c-8dfc-04274acd600f-kube-api-access-24gn2\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.936262 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.954634 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.984960 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:23Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.999333 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.999380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.999397 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.999421 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:23 crc kubenswrapper[4829]: I1002 07:17:23.999439 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:23Z","lastTransitionTime":"2025-10-02T07:17:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.003351 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.023556 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c45bd15-86f4-4d9c-8dfc-04274acd600f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.023618 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c45bd15-86f4-4d9c-8dfc-04274acd600f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.023646 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c45bd15-86f4-4d9c-8dfc-04274acd600f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.023671 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24gn2\" (UniqueName: \"kubernetes.io/projected/8c45bd15-86f4-4d9c-8dfc-04274acd600f-kube-api-access-24gn2\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.024085 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.024480 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8c45bd15-86f4-4d9c-8dfc-04274acd600f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.024748 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8c45bd15-86f4-4d9c-8dfc-04274acd600f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.031950 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8c45bd15-86f4-4d9c-8dfc-04274acd600f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.041904 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.050719 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24gn2\" (UniqueName: \"kubernetes.io/projected/8c45bd15-86f4-4d9c-8dfc-04274acd600f-kube-api-access-24gn2\") pod \"ovnkube-control-plane-749d76644c-2b4sc\" (UID: \"8c45bd15-86f4-4d9c-8dfc-04274acd600f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.059529 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.094185 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:21Z\\\",\\\"message\\\":\\\"minpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290569 6153 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290824 6153 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:21.290916 6153 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:17:21.291269 6153 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:21.291293 6153 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:21.291314 6153 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:21.291315 6153 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:21.291323 6153 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:21.291360 6153 factory.go:656] Stopping watch factory\\\\nI1002 07:17:21.291378 6153 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:21.291381 6153 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:21.291389 6153 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.103339 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.103400 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.103421 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.103446 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.103465 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.110537 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.132083 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.150332 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.160725 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.174642 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.189493 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.207365 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.207426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.207444 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.207470 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.207488 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.224372 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.244051 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.261813 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.276952 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.294105 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.310740 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.310876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.310905 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.310930 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.310947 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.323708 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c9f6ec2a2546a99d6448959c16fbb8667663f18bf56c02bbe28400a6ae3a760\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:21Z\\\",\\\"message\\\":\\\"minpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290569 6153 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:21.290824 6153 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:21.290916 6153 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 07:17:21.291269 6153 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:21.291293 6153 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:21.291314 6153 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:21.291315 6153 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:21.291323 6153 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:21.291360 6153 factory.go:656] Stopping watch factory\\\\nI1002 07:17:21.291378 6153 ovnkube.go:599] Stopped ovnkube\\\\nI1002 07:17:21.291381 6153 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:21.291389 6153 handler.go:208] Removed *v1.Node event handler 2\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.340468 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.361789 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.377858 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.397948 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.413358 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.413426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.413451 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.413520 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.413549 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.413681 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.432367 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.453007 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.469654 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.483795 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.516931 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.517013 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.517036 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.517066 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.517091 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.620079 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.620139 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.620154 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.620173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.620188 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.722385 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.722680 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.722801 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.722890 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.722966 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.825460 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.825500 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.825514 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.825532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.825544 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.881926 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/1.log" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.887357 4829 scope.go:117] "RemoveContainer" containerID="1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f" Oct 02 07:17:24 crc kubenswrapper[4829]: E1002 07:17:24.887603 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.888984 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" event={"ID":"8c45bd15-86f4-4d9c-8dfc-04274acd600f","Type":"ContainerStarted","Data":"a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.889110 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" event={"ID":"8c45bd15-86f4-4d9c-8dfc-04274acd600f","Type":"ContainerStarted","Data":"053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.889171 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" event={"ID":"8c45bd15-86f4-4d9c-8dfc-04274acd600f","Type":"ContainerStarted","Data":"8e345e616b822d85f300eaeb269d4cedc5e1597d68a2630dcd60aa50a39a8c5c"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.905610 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.924979 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.928793 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.928996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.929118 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.929213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.929334 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:24Z","lastTransitionTime":"2025-10-02T07:17:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.945585 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.963719 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.981999 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:24 crc kubenswrapper[4829]: I1002 07:17:24.999209 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:24Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.030391 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.031554 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.031707 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.031805 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.031909 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.032005 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.045506 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.066513 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.083140 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.104457 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.124384 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.135811 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.135865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.135884 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.135910 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.135960 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.146489 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.167851 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.184122 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.200519 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.219126 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.237167 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.237255 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237441 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237468 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237489 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237568 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:41.237527225 +0000 UTC m=+52.577175660 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237695 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237739 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237760 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.237839 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:41.237814353 +0000 UTC m=+52.577462798 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.238925 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.238976 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.238996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.239052 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.239071 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.239902 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.273652 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.291138 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.311552 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.323136 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.336913 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.337648 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.337836 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.337883 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:17:41.337835145 +0000 UTC m=+52.677483600 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.337988 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.338016 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.338096 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:41.338073092 +0000 UTC m=+52.677721527 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.338132 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.338284 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:41.338207246 +0000 UTC m=+52.677855711 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.341644 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.341775 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.341891 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.342000 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.342082 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.353128 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.373962 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.391072 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.403067 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.415393 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.429141 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.445154 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.445219 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.445263 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.445289 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.445306 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.460848 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.460847 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.461044 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.461030 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.461277 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.461515 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.469209 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.489996 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.510463 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.548797 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.549269 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.549411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.549542 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.549676 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.653794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.653965 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.653993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.654064 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.654091 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.730648 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-mxmxh"] Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.731498 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.731611 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.741934 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.741985 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwdx6\" (UniqueName: \"kubernetes.io/projected/675c4639-4b87-404e-8258-0e5bae51d933-kube-api-access-gwdx6\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.759032 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.759194 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.759380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.760343 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.760403 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.760423 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.779076 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.806630 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.820351 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.835460 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.843772 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.843851 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwdx6\" (UniqueName: \"kubernetes.io/projected/675c4639-4b87-404e-8258-0e5bae51d933-kube-api-access-gwdx6\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.843942 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.844042 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:26.344017964 +0000 UTC m=+37.683666399 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.852699 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.862975 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.863036 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.863054 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.863081 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.863099 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.867071 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.878558 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwdx6\" (UniqueName: \"kubernetes.io/projected/675c4639-4b87-404e-8258-0e5bae51d933-kube-api-access-gwdx6\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.890588 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.894510 4829 scope.go:117] "RemoveContainer" containerID="1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f" Oct 02 07:17:25 crc kubenswrapper[4829]: E1002 07:17:25.894806 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.915771 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.931309 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.946438 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.962708 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.966471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.966536 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.966553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.966578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.966597 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:25Z","lastTransitionTime":"2025-10-02T07:17:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:25 crc kubenswrapper[4829]: I1002 07:17:25.984418 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:25Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.010710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.010742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.010750 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.010764 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.010772 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.019923 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.027995 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.031972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.031992 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.031999 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.032011 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.032021 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.041936 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.052683 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.057732 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.057764 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.057773 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.057788 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.057798 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.059692 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.073754 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.076208 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.079937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.079992 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.080009 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.080035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.080053 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.098995 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.103388 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.103418 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.103426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.103439 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.103450 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.116448 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:26Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.116581 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.118710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.118752 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.118765 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.118782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.118796 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.221482 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.221537 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.221556 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.221579 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.221597 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.324205 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.324358 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.324381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.324412 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.324430 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.350844 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.351053 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:26 crc kubenswrapper[4829]: E1002 07:17:26.351172 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:27.351144391 +0000 UTC m=+38.690792836 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.427795 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.427849 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.427866 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.427913 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.427933 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.530828 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.530951 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.530973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.530996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.531013 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.633887 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.634275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.634450 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.634591 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.634723 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.753959 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.754015 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.754033 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.754058 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.754076 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.857678 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.857742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.857763 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.857787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.857805 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.960922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.960976 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.960995 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.961018 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:26 crc kubenswrapper[4829]: I1002 07:17:26.961037 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:26Z","lastTransitionTime":"2025-10-02T07:17:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.064266 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.064326 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.064347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.064373 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.064392 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.167785 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.167850 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.167871 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.167900 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.167924 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.271546 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.271602 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.271620 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.271645 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.271663 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.361649 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:27 crc kubenswrapper[4829]: E1002 07:17:27.361918 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:27 crc kubenswrapper[4829]: E1002 07:17:27.362054 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:29.362024645 +0000 UTC m=+40.701673080 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.374038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.374102 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.374125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.374150 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.374169 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.460643 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.460650 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.460921 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.461027 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:27 crc kubenswrapper[4829]: E1002 07:17:27.461183 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:27 crc kubenswrapper[4829]: E1002 07:17:27.461361 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:27 crc kubenswrapper[4829]: E1002 07:17:27.461532 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.461559 4829 scope.go:117] "RemoveContainer" containerID="ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621" Oct 02 07:17:27 crc kubenswrapper[4829]: E1002 07:17:27.461766 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.477313 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.477382 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.477465 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.477499 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.477526 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.581573 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.581631 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.581653 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.581679 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.581694 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.684037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.684096 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.684113 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.684138 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.684157 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.786898 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.786961 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.786980 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.787008 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.787026 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.889726 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.889785 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.889797 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.889814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.889846 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.902816 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.904617 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.905147 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.920994 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.935488 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.950496 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.965346 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.992041 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.992097 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.992114 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.992138 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.992155 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:27Z","lastTransitionTime":"2025-10-02T07:17:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:27 crc kubenswrapper[4829]: I1002 07:17:27.996762 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.022659 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.042986 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.060801 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.074779 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.090580 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.094760 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.094809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.094826 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.094852 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.094872 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.113670 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.130098 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.145269 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.164972 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.179726 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.197568 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.197626 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.197637 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.197653 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.197665 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.202537 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.215795 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:28Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.299882 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.299920 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.299932 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.299955 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.299972 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.402972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.403096 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.403163 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.403210 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.403277 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.506194 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.506281 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.506299 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.506322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.506339 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.608660 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.608717 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.608736 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.608760 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.608779 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.711445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.711484 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.711492 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.711505 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.711514 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.813604 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.813657 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.813669 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.813687 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.813699 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.916443 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.916491 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.916505 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.916523 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:28 crc kubenswrapper[4829]: I1002 07:17:28.916539 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:28Z","lastTransitionTime":"2025-10-02T07:17:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.019406 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.019463 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.019479 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.019504 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.019522 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.127598 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.127687 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.127710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.127748 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.127772 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.230735 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.230803 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.230821 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.230845 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.230862 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.333895 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.333986 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.334007 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.334033 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.334052 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.382306 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:29 crc kubenswrapper[4829]: E1002 07:17:29.382498 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:29 crc kubenswrapper[4829]: E1002 07:17:29.382576 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:33.382552427 +0000 UTC m=+44.722200872 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.437478 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.437546 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.437570 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.437598 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.437617 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.459902 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.459934 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.459984 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.460040 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:29 crc kubenswrapper[4829]: E1002 07:17:29.460045 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:29 crc kubenswrapper[4829]: E1002 07:17:29.460123 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:29 crc kubenswrapper[4829]: E1002 07:17:29.460311 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:29 crc kubenswrapper[4829]: E1002 07:17:29.460514 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.480701 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.498113 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.516973 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.539704 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.539754 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.539771 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.539794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.539812 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.548774 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.569574 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.589440 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.603119 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.619857 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.643608 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.643705 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.643755 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.643786 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.643839 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.650503 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.668377 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.687883 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.704558 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.727483 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.746775 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.746857 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.746883 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.746915 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.746940 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.748693 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.781385 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.803405 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.820392 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.849568 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.849629 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.849647 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.849675 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.849694 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.952721 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.952796 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.952814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.952841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:29 crc kubenswrapper[4829]: I1002 07:17:29.952859 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:29Z","lastTransitionTime":"2025-10-02T07:17:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.055495 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.055567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.055585 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.055611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.055634 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.158743 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.158808 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.158825 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.158856 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.158875 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.262307 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.262390 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.262471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.262504 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.262527 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.365205 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.365284 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.365296 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.365312 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.365322 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.468019 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.468072 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.468088 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.468110 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.468123 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.571184 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.571290 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.571315 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.571347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.571368 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.674840 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.674906 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.674930 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.674958 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.674980 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.778747 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.778824 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.778849 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.778876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.778894 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.882577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.882652 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.882680 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.882709 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.882732 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.986150 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.986219 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.986279 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.986314 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:30 crc kubenswrapper[4829]: I1002 07:17:30.986342 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:30Z","lastTransitionTime":"2025-10-02T07:17:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.089356 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.089433 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.089462 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.089497 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.089524 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.192200 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.192282 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.192304 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.192326 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.192343 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.295710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.295769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.295788 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.295812 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.295831 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.398834 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.398916 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.398941 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.398974 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.398992 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.463368 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.463480 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.463519 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:31 crc kubenswrapper[4829]: E1002 07:17:31.463561 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.463640 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:31 crc kubenswrapper[4829]: E1002 07:17:31.463834 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:31 crc kubenswrapper[4829]: E1002 07:17:31.463960 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:31 crc kubenswrapper[4829]: E1002 07:17:31.464050 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.501781 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.501836 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.501854 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.501879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.501900 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.605646 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.605719 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.605740 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.605811 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.605914 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.709040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.709087 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.709156 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.709181 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.709198 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.812560 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.812629 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.812651 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.812681 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.812703 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.915425 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.915485 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.915501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.915526 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:31 crc kubenswrapper[4829]: I1002 07:17:31.915542 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:31Z","lastTransitionTime":"2025-10-02T07:17:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.019434 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.019500 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.019517 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.019547 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.019564 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.122934 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.123853 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.123990 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.124124 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.124299 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.227918 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.228322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.228505 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.228668 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.228809 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.331974 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.332042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.332073 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.332123 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.332153 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.435251 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.435323 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.435342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.435379 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.435404 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.538723 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.538787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.538810 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.538837 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.538856 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.642336 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.642393 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.642412 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.642437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.642456 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.746031 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.746115 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.746141 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.746176 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.746200 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.850193 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.850296 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.850319 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.850350 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.850374 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.953681 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.953744 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.953769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.953801 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:32 crc kubenswrapper[4829]: I1002 07:17:32.953823 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:32Z","lastTransitionTime":"2025-10-02T07:17:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.056920 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.056960 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.056973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.056991 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.057003 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.160489 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.160558 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.160578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.160602 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.160619 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.263911 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.263989 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.264015 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.264048 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.264074 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.367735 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.367789 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.367811 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.367841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.367864 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.428903 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:33 crc kubenswrapper[4829]: E1002 07:17:33.429129 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:33 crc kubenswrapper[4829]: E1002 07:17:33.429279 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:41.429214497 +0000 UTC m=+52.768862942 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.460664 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.460730 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.460735 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:33 crc kubenswrapper[4829]: E1002 07:17:33.461305 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.460834 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:33 crc kubenswrapper[4829]: E1002 07:17:33.461438 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:33 crc kubenswrapper[4829]: E1002 07:17:33.461656 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:33 crc kubenswrapper[4829]: E1002 07:17:33.461797 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.473048 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.473119 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.473143 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.473173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.473196 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.578209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.578438 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.578516 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.578558 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.579402 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.681125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.681163 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.681171 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.681186 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.681195 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.783999 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.784285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.784368 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.784465 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.784539 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.888152 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.888219 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.888264 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.888290 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.888307 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.991042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.991101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.991120 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.991144 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:33 crc kubenswrapper[4829]: I1002 07:17:33.991160 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:33Z","lastTransitionTime":"2025-10-02T07:17:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.094294 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.094396 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.094416 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.094440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.094457 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.197768 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.197841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.197866 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.197896 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.197915 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.300909 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.301497 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.301653 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.301771 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.301888 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.404459 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.404523 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.404535 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.404551 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.404563 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.507456 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.507544 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.507556 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.507602 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.507614 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.611402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.611776 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.611936 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.612083 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.612250 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.720364 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.720427 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.720450 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.720483 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.720505 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.823818 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.823922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.823942 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.823967 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.823985 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.929054 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.929119 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.929137 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.929162 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:34 crc kubenswrapper[4829]: I1002 07:17:34.929179 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:34Z","lastTransitionTime":"2025-10-02T07:17:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.032353 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.032422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.032439 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.032464 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.032482 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.135839 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.135907 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.135925 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.135952 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.135970 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.239216 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.239312 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.239330 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.239356 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.239374 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.342215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.342306 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.342326 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.342352 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.342371 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.449449 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.449810 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.450073 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.450259 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.450445 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.460310 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.460363 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.460413 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:35 crc kubenswrapper[4829]: E1002 07:17:35.460824 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.460426 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:35 crc kubenswrapper[4829]: E1002 07:17:35.461020 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:35 crc kubenswrapper[4829]: E1002 07:17:35.461173 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:35 crc kubenswrapper[4829]: E1002 07:17:35.461474 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.554979 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.555035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.555054 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.555078 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.555097 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.658398 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.658458 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.658476 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.658508 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.658527 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.761767 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.761811 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.761819 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.761835 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.761846 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.864703 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.864767 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.864783 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.864805 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.864823 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.968074 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.968122 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.968134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.968153 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:35 crc kubenswrapper[4829]: I1002 07:17:35.968165 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:35Z","lastTransitionTime":"2025-10-02T07:17:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.071364 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.071441 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.071466 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.071493 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.071510 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.174738 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.174779 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.174787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.174808 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.174818 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.290288 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.290342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.290350 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.290365 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.290375 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.393108 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.393285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.393313 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.393344 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.393368 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.418086 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.418156 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.418175 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.418201 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.418221 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: E1002 07:17:36.438567 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.442412 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.442461 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.442479 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.442504 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.442520 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.462555 4829 scope.go:117] "RemoveContainer" containerID="1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f" Oct 02 07:17:36 crc kubenswrapper[4829]: E1002 07:17:36.463724 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.468730 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.468774 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.468792 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.468813 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.468828 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: E1002 07:17:36.490176 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.495551 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.495603 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.495618 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.495642 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.495658 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: E1002 07:17:36.521132 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.526009 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.526058 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.526079 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.526108 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.526132 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: E1002 07:17:36.544926 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4829]: E1002 07:17:36.545190 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.547379 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.547432 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.547452 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.547477 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.547494 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.650729 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.650772 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.650782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.650799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.650808 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.758587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.758643 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.758658 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.758681 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.758696 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.862050 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.862111 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.862127 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.862150 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.862168 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.940155 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/1.log" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.945400 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.946210 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.965606 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.965663 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.965677 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.965701 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.965721 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:36Z","lastTransitionTime":"2025-10-02T07:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.976722 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:36 crc kubenswrapper[4829]: I1002 07:17:36.999570 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:36Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.029861 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.045293 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.057382 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.068198 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.068274 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.068292 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.068318 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.068333 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.073985 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.090194 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.106465 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.124354 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.141393 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.164163 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.170948 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.170989 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.171002 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.171021 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.171034 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.180523 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.196155 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.210589 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.226159 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.238251 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.250037 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.273520 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.273565 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.273577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.273596 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.273610 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.375994 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.376051 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.376068 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.376093 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.376113 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.460544 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.460594 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.460671 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.460544 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:37 crc kubenswrapper[4829]: E1002 07:17:37.460742 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:37 crc kubenswrapper[4829]: E1002 07:17:37.460914 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:37 crc kubenswrapper[4829]: E1002 07:17:37.460980 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:37 crc kubenswrapper[4829]: E1002 07:17:37.461059 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.478935 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.478996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.479016 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.479044 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.479068 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.582582 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.582645 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.582664 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.582695 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.582764 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.685863 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.685928 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.685946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.685970 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.685988 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.789772 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.789834 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.789853 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.789879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.789897 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.893302 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.893367 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.893383 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.893409 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.893427 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.952990 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/2.log" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.953879 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/1.log" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.958650 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40" exitCode=1 Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.958705 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40"} Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.958753 4829 scope.go:117] "RemoveContainer" containerID="1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.960285 4829 scope.go:117] "RemoveContainer" containerID="574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40" Oct 02 07:17:37 crc kubenswrapper[4829]: E1002 07:17:37.960584 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.984668 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:37Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.996274 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.996342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.996362 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.996388 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:37 crc kubenswrapper[4829]: I1002 07:17:37.996411 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:37Z","lastTransitionTime":"2025-10-02T07:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.008820 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.031799 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.053345 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.072008 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.088793 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.099175 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.099255 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.099272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.099297 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.099315 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.107005 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.131882 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.164656 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.201565 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.201606 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.201620 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.201637 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.201651 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.207138 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.220416 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.242199 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.255723 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.282274 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1755fc8077ce003e771a82c0ab7efc2ab26562e120c2b267d7e4877df759125f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"message\\\":\\\" 6287 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1002 07:17:23.100391 6287 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 07:17:23.100415 6287 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 07:17:23.100426 6287 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 07:17:23.100430 6287 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:23.100440 6287 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:23.100466 6287 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 07:17:23.100472 6287 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 07:17:23.100496 6287 factory.go:656] Stopping watch factory\\\\nI1002 07:17:23.100508 6287 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:23.100515 6287 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:23.100521 6287 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:23.100517 6287 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 07:17:23.100528 6287 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:23.100529 6287 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:23.100561 6287 handler.go:208] Removed *v1.Node ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.291943 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.302122 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.303777 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.303799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.303807 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.303819 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.303828 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.313480 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.406394 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.406448 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.406465 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.406492 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.406511 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.510119 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.510173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.510190 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.510217 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.510263 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.613257 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.613329 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.613352 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.613382 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.613405 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.715987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.716038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.716060 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.716084 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.716100 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.818315 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.818381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.818401 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.818422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.818441 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.922673 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.922732 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.922796 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.922824 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.922844 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:38Z","lastTransitionTime":"2025-10-02T07:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.964886 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/2.log" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.969896 4829 scope.go:117] "RemoveContainer" containerID="574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40" Oct 02 07:17:38 crc kubenswrapper[4829]: E1002 07:17:38.970171 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:17:38 crc kubenswrapper[4829]: I1002 07:17:38.994112 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:38Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.014967 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.025726 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.025799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.025823 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.025847 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.025864 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.048295 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.066986 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.084097 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.108839 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.128979 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.129045 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.129065 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.129090 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.129110 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.129728 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.155441 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.177544 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.197944 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.215885 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.231403 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.231455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.231474 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.231498 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.231519 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.232934 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.250475 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.283763 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.303996 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.323699 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.335149 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.335215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.335268 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.335295 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.335314 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.342328 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.438377 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.438459 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.438479 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.438505 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.438523 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.459846 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.459894 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.459912 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.460017 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:39 crc kubenswrapper[4829]: E1002 07:17:39.460134 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:39 crc kubenswrapper[4829]: E1002 07:17:39.460329 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:39 crc kubenswrapper[4829]: E1002 07:17:39.460505 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:39 crc kubenswrapper[4829]: E1002 07:17:39.460613 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.478485 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.495649 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.516666 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.540577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.540630 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.540663 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.540710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.540729 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.542890 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.562760 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.582073 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.599560 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.632025 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.648114 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.648218 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.648275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.648307 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.648342 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.654896 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.667421 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.680497 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.700964 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.717990 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.741207 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.751885 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.751934 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.751953 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.751978 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.751995 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.767297 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.789128 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.806554 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:39Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.854423 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.854491 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.854509 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.854535 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.854554 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.958629 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.958699 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.958717 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.958742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:39 crc kubenswrapper[4829]: I1002 07:17:39.958764 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:39Z","lastTransitionTime":"2025-10-02T07:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.061758 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.061827 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.061846 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.061871 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.061887 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.165529 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.165601 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.165620 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.165648 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.165667 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.268641 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.268705 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.268730 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.268756 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.268772 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.371671 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.371724 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.371744 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.371769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.371786 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.474932 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.474997 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.475014 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.475038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.475055 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.578207 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.578294 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.578312 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.578336 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.578353 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.681844 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.681927 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.681945 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.681974 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.681994 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.785567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.785630 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.785647 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.785670 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.785687 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.888492 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.888547 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.888576 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.888597 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.888612 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.992285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.992353 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.992376 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.992402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:40 crc kubenswrapper[4829]: I1002 07:17:40.992422 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:40Z","lastTransitionTime":"2025-10-02T07:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.095838 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.095904 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.095922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.095950 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.095967 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.198867 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.198932 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.198957 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.198986 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.199008 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.242841 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.243138 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.243466 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.243500 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.243523 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.243640 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:13.243614742 +0000 UTC m=+84.583263187 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.244604 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.244644 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.244664 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.244733 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:13.244702894 +0000 UTC m=+84.584351339 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.302120 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.302162 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.302173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.302189 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.302201 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.343766 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.344020 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:18:13.343977674 +0000 UTC m=+84.683626119 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.344080 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.344165 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.344345 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.344367 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.344429 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:13.344406486 +0000 UTC m=+84.684054931 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.344485 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:13.344461439 +0000 UTC m=+84.684109884 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.405519 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.405584 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.405602 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.405631 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.405655 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.445764 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.445983 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.446094 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:17:57.446065087 +0000 UTC m=+68.785713522 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.460151 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.460251 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.460155 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.460381 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.460392 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.460554 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.460721 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:41 crc kubenswrapper[4829]: E1002 07:17:41.460909 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.508530 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.508587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.508604 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.508627 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.508643 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.611723 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.611823 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.611846 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.611875 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.611897 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.714498 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.714552 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.714570 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.714652 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.714679 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.817595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.817672 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.817696 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.817727 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.817750 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.921209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.921316 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.921343 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.921376 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:41 crc kubenswrapper[4829]: I1002 07:17:41.921402 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:41Z","lastTransitionTime":"2025-10-02T07:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.024301 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.024363 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.024377 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.024398 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.024411 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.127497 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.127830 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.127846 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.127874 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.127892 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.230691 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.230764 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.230816 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.230847 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.230869 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.334471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.334606 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.334688 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.334804 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.334833 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.438437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.438505 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.438527 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.438557 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.438576 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.540886 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.540928 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.540939 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.540958 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.540970 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.643638 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.643726 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.643749 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.643781 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.643805 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.746926 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.746987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.747003 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.747031 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.747048 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.850171 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.850270 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.850288 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.850314 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.850365 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.953583 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.953650 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.953671 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.953695 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:42 crc kubenswrapper[4829]: I1002 07:17:42.953711 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:42Z","lastTransitionTime":"2025-10-02T07:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.056205 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.056321 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.056341 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.056365 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.056383 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.159541 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.159595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.159611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.159634 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.159654 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.261468 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.261504 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.261514 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.261528 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.261539 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.363733 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.363782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.363792 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.363809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.363817 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.460340 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.460391 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.460438 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:43 crc kubenswrapper[4829]: E1002 07:17:43.460657 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.461212 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:43 crc kubenswrapper[4829]: E1002 07:17:43.461404 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:43 crc kubenswrapper[4829]: E1002 07:17:43.461524 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:43 crc kubenswrapper[4829]: E1002 07:17:43.461633 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.466387 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.466575 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.466708 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.466835 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.466985 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.570351 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.570409 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.570428 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.570451 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.570467 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.672943 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.673004 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.673023 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.673045 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.673061 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.775698 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.775755 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.775780 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.775808 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.775827 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.878686 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.878734 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.878752 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.878775 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.878791 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.972164 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.981862 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.981922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.981947 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.981976 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.981997 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:43Z","lastTransitionTime":"2025-10-02T07:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:43 crc kubenswrapper[4829]: I1002 07:17:43.993320 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:43Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.012590 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.042493 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.058355 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.074855 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.084992 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.085049 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.085068 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.085092 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.085109 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.099130 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.118458 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.142288 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.166612 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.187628 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.187702 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.187665 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.187725 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.187972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.187998 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.206769 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.224553 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.241015 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.264201 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.284097 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.290711 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.290859 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.290963 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.291088 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.291208 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.298528 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.312500 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.393787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.393855 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.393873 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.393901 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.393918 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.496759 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.496814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.496840 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.496864 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.496881 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.507308 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.521471 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.529807 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.548164 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.567928 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.584214 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.599517 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.599794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.599905 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.600003 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.600103 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.605786 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.635504 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.655912 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.676142 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.697698 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.702490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.706659 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.706889 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.707034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.707172 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.718441 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.739918 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.769978 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.784179 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.798809 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.810339 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.810565 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.810703 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.810876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.811033 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.814837 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.831808 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.855584 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:44Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.914309 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.914381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.914401 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.914426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:44 crc kubenswrapper[4829]: I1002 07:17:44.914449 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:44Z","lastTransitionTime":"2025-10-02T07:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.016894 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.016972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.016998 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.017030 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.017050 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.120828 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.120896 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.120913 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.120939 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.120957 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.223806 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.223862 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.223879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.223904 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.223920 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.327177 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.327267 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.327287 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.327325 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.327343 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.430740 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.430796 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.430814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.430838 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.430856 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.460358 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:45 crc kubenswrapper[4829]: E1002 07:17:45.460574 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.460640 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.460682 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.460640 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:45 crc kubenswrapper[4829]: E1002 07:17:45.460873 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:45 crc kubenswrapper[4829]: E1002 07:17:45.460989 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:45 crc kubenswrapper[4829]: E1002 07:17:45.461080 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.533891 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.533970 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.533987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.534009 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.534028 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.637336 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.637389 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.637408 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.637431 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.637448 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.740875 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.740922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.740938 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.740961 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.740978 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.845423 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.845486 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.845503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.845529 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.845547 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.948724 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.948792 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.948814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.948843 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:45 crc kubenswrapper[4829]: I1002 07:17:45.948864 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:45Z","lastTransitionTime":"2025-10-02T07:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.051670 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.051747 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.051771 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.051809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.051833 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.154739 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.154817 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.154843 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.154876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.154899 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.257743 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.257816 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.257834 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.257858 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.257878 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.361367 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.361440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.361463 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.361488 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.361510 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.465064 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.465155 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.465177 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.465207 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.465261 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.569052 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.569118 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.569134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.569158 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.569174 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.672866 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.672952 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.672981 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.673012 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.673048 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.779380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.779456 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.779475 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.779502 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.779519 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: E1002 07:17:46.799291 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.806586 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.806674 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.806693 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.806719 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.806737 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: E1002 07:17:46.827140 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.832399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.832464 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.832487 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.832511 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.832531 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: E1002 07:17:46.852827 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.858296 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.858360 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.858381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.858411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.858432 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: E1002 07:17:46.879535 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.884844 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.884919 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.884944 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.884975 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.884998 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:46 crc kubenswrapper[4829]: E1002 07:17:46.904898 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:46Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:46 crc kubenswrapper[4829]: E1002 07:17:46.905125 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.907545 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.907623 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.907641 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.907667 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:46 crc kubenswrapper[4829]: I1002 07:17:46.907685 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:46Z","lastTransitionTime":"2025-10-02T07:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.010788 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.010845 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.010866 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.010891 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.010911 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.114128 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.114192 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.114210 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.114260 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.114279 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.216835 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.216902 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.216938 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.216970 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.216991 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.321074 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.321140 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.321161 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.321192 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.321214 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.423687 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.423742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.423759 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.423783 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.423801 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.460954 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.461053 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:47 crc kubenswrapper[4829]: E1002 07:17:47.461177 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.461270 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.460957 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:47 crc kubenswrapper[4829]: E1002 07:17:47.461409 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:47 crc kubenswrapper[4829]: E1002 07:17:47.461532 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:47 crc kubenswrapper[4829]: E1002 07:17:47.461643 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.526625 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.526670 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.526681 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.526698 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.526712 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.629510 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.629570 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.629587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.629609 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.629626 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.732306 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.732367 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.732386 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.732408 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.732425 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.835920 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.835996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.836022 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.836054 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.836080 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.938559 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.938611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.938627 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.938649 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:47 crc kubenswrapper[4829]: I1002 07:17:47.938666 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:47Z","lastTransitionTime":"2025-10-02T07:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.041309 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.041374 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.041395 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.041421 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.041438 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.145327 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.145401 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.145419 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.145444 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.145460 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.249539 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.249603 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.249616 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.249639 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.249653 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.352984 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.353040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.353056 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.353080 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.353098 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.455743 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.455806 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.455823 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.455847 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.455865 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.559867 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.559961 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.559987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.560021 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.560048 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.663452 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.663514 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.663533 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.663614 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.663635 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.766575 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.766656 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.766679 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.766703 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.766720 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.869459 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.869520 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.869541 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.869566 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.869584 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.972946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.973009 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.973027 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.973052 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:48 crc kubenswrapper[4829]: I1002 07:17:48.973069 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:48Z","lastTransitionTime":"2025-10-02T07:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.076819 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.076892 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.076917 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.076949 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.076970 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.180672 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.180755 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.180774 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.180799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.180819 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.284115 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.284293 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.284325 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.284361 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.284383 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.388411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.388473 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.388495 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.388527 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.388552 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.459803 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.459885 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.459896 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.459974 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:49 crc kubenswrapper[4829]: E1002 07:17:49.460035 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:49 crc kubenswrapper[4829]: E1002 07:17:49.460175 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:49 crc kubenswrapper[4829]: E1002 07:17:49.460350 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:49 crc kubenswrapper[4829]: E1002 07:17:49.461448 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.480444 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.493464 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.493524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.493548 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.493578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.493600 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.500590 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.538328 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.555513 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.571377 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.586713 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.597026 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.597081 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.597101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.597126 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.597143 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.601801 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.621154 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.635869 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.652752 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.672659 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.688445 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.700003 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.700285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.700815 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.701144 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.701520 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.705930 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.739402 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.758716 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.780750 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.803730 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.804751 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.804825 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.804843 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.804866 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.804914 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.823147 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:49Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.908291 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.908342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.908358 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.908381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:49 crc kubenswrapper[4829]: I1002 07:17:49.908401 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:49Z","lastTransitionTime":"2025-10-02T07:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.011134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.011191 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.011208 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.011256 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.011276 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.114730 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.114865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.114891 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.114920 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.114944 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.218135 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.218196 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.218213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.218272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.218300 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.321492 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.321567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.321585 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.321609 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.321626 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.424551 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.424827 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.424849 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.424873 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.424893 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.528812 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.528888 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.528910 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.528939 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.528962 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.632275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.632346 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.632370 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.632399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.632419 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.735678 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.735742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.735759 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.735783 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.735800 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.839040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.839093 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.839110 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.839134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.839151 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.942285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.942338 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.942354 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.942382 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:50 crc kubenswrapper[4829]: I1002 07:17:50.942400 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:50Z","lastTransitionTime":"2025-10-02T07:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.045680 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.045741 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.045765 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.045793 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.045813 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.148966 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.149034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.149056 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.149087 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.149110 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.252598 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.252674 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.252697 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.252732 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.252758 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.355845 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.355919 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.355938 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.355964 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.355982 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.459699 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.459750 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.459769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.459793 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.459809 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.460484 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.460540 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.460484 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:51 crc kubenswrapper[4829]: E1002 07:17:51.460701 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.460722 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:51 crc kubenswrapper[4829]: E1002 07:17:51.460861 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:51 crc kubenswrapper[4829]: E1002 07:17:51.460986 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:51 crc kubenswrapper[4829]: E1002 07:17:51.461155 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.563463 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.563520 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.563533 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.563555 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.563570 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.666037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.666089 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.666105 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.666126 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.666141 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.768336 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.768388 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.768412 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.768439 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.768459 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.872075 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.872182 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.872204 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.872275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.872315 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.975716 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.975820 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.975846 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.975876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:51 crc kubenswrapper[4829]: I1002 07:17:51.975898 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:51Z","lastTransitionTime":"2025-10-02T07:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.079004 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.079122 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.079145 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.079176 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.079197 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.182991 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.183606 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.183630 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.183666 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.183689 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.286913 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.287000 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.287019 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.287043 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.287061 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.389486 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.389535 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.389551 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.389572 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.389588 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.461433 4829 scope.go:117] "RemoveContainer" containerID="574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40" Oct 02 07:17:52 crc kubenswrapper[4829]: E1002 07:17:52.461787 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.492726 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.492782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.492805 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.492830 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.492851 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.596293 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.596352 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.596369 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.596394 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.596415 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.699034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.699072 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.699113 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.699133 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.699145 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.801964 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.802002 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.802039 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.802059 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.802072 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.904706 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.904782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.904802 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.904833 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:52 crc kubenswrapper[4829]: I1002 07:17:52.904853 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:52Z","lastTransitionTime":"2025-10-02T07:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.008091 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.008146 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.008159 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.008179 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.008198 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.110486 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.110553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.110571 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.110597 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.110615 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.214314 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.214394 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.214417 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.214445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.214465 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.316803 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.316892 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.316917 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.316948 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.316974 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.420040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.420096 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.420113 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.420138 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.420155 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.459955 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.459994 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.459955 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:53 crc kubenswrapper[4829]: E1002 07:17:53.460106 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:53 crc kubenswrapper[4829]: E1002 07:17:53.460413 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:53 crc kubenswrapper[4829]: E1002 07:17:53.460611 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.460671 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:53 crc kubenswrapper[4829]: E1002 07:17:53.460763 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.523524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.523586 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.523597 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.523619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.523632 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.626747 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.626809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.626827 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.626852 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.626872 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.729827 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.729878 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.729896 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.729920 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.729938 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.833402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.833462 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.833479 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.833503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.833520 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.936471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.936525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.936542 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.936565 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:53 crc kubenswrapper[4829]: I1002 07:17:53.936585 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:53Z","lastTransitionTime":"2025-10-02T07:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.038215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.038273 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.038285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.038300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.038313 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.140324 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.140373 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.140389 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.140411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.140429 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.242973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.243205 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.243286 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.243381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.243450 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.345275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.345310 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.345320 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.345337 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.345348 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.447162 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.447199 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.447210 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.447239 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.447251 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.548817 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.548887 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.548898 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.548918 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.548931 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.651396 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.651467 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.651483 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.651502 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.651514 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.754396 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.754484 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.754509 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.754543 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.754570 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.857271 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.857335 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.857353 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.857377 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.857394 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.960291 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.960357 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.960380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.960405 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:54 crc kubenswrapper[4829]: I1002 07:17:54.960423 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:54Z","lastTransitionTime":"2025-10-02T07:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.062087 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.062115 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.062125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.062141 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.062151 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.164471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.164503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.164514 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.164528 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.164539 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.266986 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.267011 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.267022 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.267034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.267045 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.370167 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.370300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.370328 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.370357 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.370380 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.460297 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.460370 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:55 crc kubenswrapper[4829]: E1002 07:17:55.460404 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.460297 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.460302 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:55 crc kubenswrapper[4829]: E1002 07:17:55.460588 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:55 crc kubenswrapper[4829]: E1002 07:17:55.460709 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:55 crc kubenswrapper[4829]: E1002 07:17:55.460755 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.472380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.472413 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.472423 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.472437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.472449 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.574809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.574842 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.574850 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.574863 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.574871 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.677549 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.677573 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.677581 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.677592 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.677600 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.780126 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.780154 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.780164 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.780174 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.780183 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.882878 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.882899 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.882909 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.882920 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.882927 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.985521 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.985578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.985596 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.985626 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:55 crc kubenswrapper[4829]: I1002 07:17:55.985643 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:55Z","lastTransitionTime":"2025-10-02T07:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.088423 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.088451 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.088458 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.088470 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.088479 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.190490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.190540 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.190556 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.190578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.190594 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.293923 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.293993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.294012 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.294037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.294054 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.396825 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.397460 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.397711 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.397947 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.398201 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.501614 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.501930 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.502057 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.502176 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.502325 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.605257 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.605315 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.605334 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.605360 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.605379 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.707195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.707289 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.707300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.707317 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.707330 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.809274 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.809321 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.809334 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.809351 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.809362 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.912219 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.912286 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.912297 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.912316 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.912330 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.928857 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.928948 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.928967 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.928993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.929013 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: E1002 07:17:56.946484 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.951351 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.951404 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.951415 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.951432 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.951442 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: E1002 07:17:56.969129 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.974591 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.974652 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.974671 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.974699 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.974717 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:56 crc kubenswrapper[4829]: E1002 07:17:56.993184 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:56Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.998821 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.998875 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.998894 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.998923 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:56 crc kubenswrapper[4829]: I1002 07:17:56.998941 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:56Z","lastTransitionTime":"2025-10-02T07:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.013923 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:57Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.019183 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.019248 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.019262 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.019284 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.019298 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.039173 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:57Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.039434 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.042794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.042845 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.042878 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.042904 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.042924 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.146828 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.147177 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.147352 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.147521 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.147649 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.250472 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.250515 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.250526 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.250542 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.250552 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.353182 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.353296 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.353318 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.353343 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.353361 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.484173 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.484458 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.484821 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.484968 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.484975 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.485191 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.485388 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.485444 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.486261 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.486341 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:57 crc kubenswrapper[4829]: E1002 07:17:57.486935 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:18:29.486898246 +0000 UTC m=+100.826546691 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.486592 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.487371 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.487663 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.487988 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.488313 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.592617 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.592671 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.592682 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.592706 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.592719 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.695724 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.695779 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.695799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.695824 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.695842 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.798874 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.798926 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.798947 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.798976 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.798996 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.906721 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.906792 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.906815 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.906843 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:57 crc kubenswrapper[4829]: I1002 07:17:57.906863 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:57Z","lastTransitionTime":"2025-10-02T07:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.010461 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.010517 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.010533 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.010559 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.010578 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.113972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.114042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.114060 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.114086 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.114103 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.216878 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.216923 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.216935 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.216950 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.216959 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.319888 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.319929 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.319939 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.319952 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.319964 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.423142 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.423449 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.423492 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.423524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.423543 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.526042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.526085 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.526098 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.526114 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.526123 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.628496 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.628555 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.628578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.628607 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.628630 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.730474 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.730503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.730511 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.730524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.730533 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.833219 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.833272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.833285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.833300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.833309 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.935086 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.935125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.935137 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.935153 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:58 crc kubenswrapper[4829]: I1002 07:17:58.935188 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:58Z","lastTransitionTime":"2025-10-02T07:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038570 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/0.log" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038615 4829 generic.go:334] "Generic (PLEG): container finished" podID="ca00dadf-1664-466b-830c-e172857db47a" containerID="983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428" exitCode=1 Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038646 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerDied","Data":"983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038708 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038748 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038764 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038791 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038807 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.038995 4829 scope.go:117] "RemoveContainer" containerID="983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.058651 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.075981 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.096837 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.116621 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.127676 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.142387 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.142362 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.142442 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.142459 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.142480 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.142497 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.153049 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.170060 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.182763 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.198195 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.211411 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.223172 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.235475 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.245793 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.245836 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.245851 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.245869 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.245885 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.246356 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.276197 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.299166 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.318815 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.335421 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.349161 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.349216 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.349247 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.349268 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.349282 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.453009 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.453080 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.453100 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.453130 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.453147 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.460385 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.460468 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.460507 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:17:59 crc kubenswrapper[4829]: E1002 07:17:59.460596 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:17:59 crc kubenswrapper[4829]: E1002 07:17:59.461147 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:17:59 crc kubenswrapper[4829]: E1002 07:17:59.461322 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.461534 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:17:59 crc kubenswrapper[4829]: E1002 07:17:59.461701 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.479066 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.490724 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.505386 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.527199 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.552485 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.556395 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.556449 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.556468 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.556491 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.556509 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.567728 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.582600 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.595371 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.605351 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.615873 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.631431 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.646446 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.659202 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.659265 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.659280 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.659300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.659312 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.663477 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.676668 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:59Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.688387 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.707094 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.719665 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.734549 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:17:59Z is after 2025-08-24T17:21:41Z" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.761534 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.761567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.761575 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.761589 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.761597 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.864995 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.865037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.865053 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.865067 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.865076 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.967200 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.967726 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.967753 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.967785 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:17:59 crc kubenswrapper[4829]: I1002 07:17:59.967808 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:17:59Z","lastTransitionTime":"2025-10-02T07:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.045496 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/0.log" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.045580 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerStarted","Data":"3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.060425 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.070769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.070842 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.070863 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.070887 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.070905 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.075975 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.101989 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.113665 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.125480 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.143713 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.155254 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.169457 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.173654 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.173703 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.173717 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.173737 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.173751 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.184305 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.199787 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.219058 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.234541 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.250819 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.275590 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.276650 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.276708 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.276725 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.276749 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.276765 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.291083 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.305199 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.325522 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.338443 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:00Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.379411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.379452 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.379462 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.379480 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.379500 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.480993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.481071 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.481085 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.481099 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.481110 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.584048 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.584085 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.584097 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.584110 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.584120 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.686716 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.686774 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.686793 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.686818 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.686836 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.789386 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.789418 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.789427 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.789440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.789450 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.891625 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.891656 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.891665 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.891679 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.891688 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.993521 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.993547 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.993557 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.993570 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:00 crc kubenswrapper[4829]: I1002 07:18:00.993578 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:00Z","lastTransitionTime":"2025-10-02T07:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.095945 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.095983 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.095999 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.096021 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.096056 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.198479 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.198535 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.198555 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.198579 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.198597 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.300665 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.300692 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.300700 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.300713 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.300720 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.402901 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.402937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.402949 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.402965 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.402978 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.460248 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.460264 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.460269 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.460242 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:01 crc kubenswrapper[4829]: E1002 07:18:01.460362 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:01 crc kubenswrapper[4829]: E1002 07:18:01.460425 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:01 crc kubenswrapper[4829]: E1002 07:18:01.460560 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:01 crc kubenswrapper[4829]: E1002 07:18:01.460710 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.505111 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.505155 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.505167 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.505182 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.505193 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.607481 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.607510 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.607518 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.607531 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.607540 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.710278 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.710319 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.710328 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.710344 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.710355 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.812977 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.813025 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.813035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.813050 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.813062 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.915175 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.915255 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.915273 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.915297 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:01 crc kubenswrapper[4829]: I1002 07:18:01.915314 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:01Z","lastTransitionTime":"2025-10-02T07:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.017591 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.017669 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.017686 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.017714 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.017733 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.122272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.122329 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.122341 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.122361 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.122372 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.225903 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.225973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.225997 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.226029 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.226049 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.329915 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.330039 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.330607 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.330704 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.330931 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.433534 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.433619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.433640 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.433672 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.433704 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.537135 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.537209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.537268 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.537301 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.537331 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.640507 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.640569 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.640587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.640614 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.640632 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.744958 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.745033 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.745062 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.745101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.745126 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.848710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.848768 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.848788 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.848817 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.848838 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.952350 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.952397 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.952406 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.952424 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:02 crc kubenswrapper[4829]: I1002 07:18:02.952436 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:02Z","lastTransitionTime":"2025-10-02T07:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.056164 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.056202 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.056211 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.056240 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.056250 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.158901 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.158948 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.158962 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.158981 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.158993 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.261722 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.261776 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.261790 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.261809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.261822 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.364858 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.364911 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.364921 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.364938 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.364948 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.460933 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.460991 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.460950 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:03 crc kubenswrapper[4829]: E1002 07:18:03.461129 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.461188 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:03 crc kubenswrapper[4829]: E1002 07:18:03.461401 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:03 crc kubenswrapper[4829]: E1002 07:18:03.461438 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:03 crc kubenswrapper[4829]: E1002 07:18:03.461509 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.467718 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.467769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.467780 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.467799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.467814 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.570099 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.570158 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.570171 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.570189 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.570201 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.673347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.673419 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.673440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.673471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.673490 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.776650 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.776710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.776723 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.776746 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.776758 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.880944 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.881017 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.881035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.881060 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.881082 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.984115 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.984288 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.984318 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.984356 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:03 crc kubenswrapper[4829]: I1002 07:18:03.984386 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:03Z","lastTransitionTime":"2025-10-02T07:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.087836 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.087887 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.087904 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.087927 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.087942 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.191725 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.191772 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.191785 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.191801 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.191813 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.295107 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.295185 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.295201 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.295255 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.295278 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.398716 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.398770 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.398787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.398809 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.398829 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.461839 4829 scope.go:117] "RemoveContainer" containerID="574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.502446 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.502508 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.502576 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.502609 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.502631 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.605587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.605657 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.605680 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.605708 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.605729 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.708500 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.708560 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.708580 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.708605 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.708624 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.811981 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.812045 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.812064 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.812088 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.812105 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.914836 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.914896 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.914913 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.914937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:04 crc kubenswrapper[4829]: I1002 07:18:04.914954 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:04Z","lastTransitionTime":"2025-10-02T07:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.017490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.017540 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.017552 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.017568 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.017578 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.068692 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/2.log" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.072993 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.073852 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.108085 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.122313 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.122387 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.122410 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.122441 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.122478 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.128387 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.155874 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.175402 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.196264 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.214909 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.224360 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.224410 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.224425 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.224446 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.224462 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.240077 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.262965 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:18:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.278973 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.290332 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.308554 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.327546 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.327598 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.327611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.327633 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.327647 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.328581 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.347019 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.363935 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.378490 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.393627 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.403639 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.430337 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.430391 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.430403 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.430422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.430434 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.439084 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.460607 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.460665 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.460640 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.460605 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:05 crc kubenswrapper[4829]: E1002 07:18:05.460788 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:05 crc kubenswrapper[4829]: E1002 07:18:05.460875 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:05 crc kubenswrapper[4829]: E1002 07:18:05.460948 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:05 crc kubenswrapper[4829]: E1002 07:18:05.461067 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.533365 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.533405 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.533414 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.533430 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.533440 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.636455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.636523 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.636542 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.636568 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.636590 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.739338 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.739403 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.739423 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.739448 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.739466 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.842327 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.842391 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.842413 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.842440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.842456 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.946279 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.946383 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.946425 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.946455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:05 crc kubenswrapper[4829]: I1002 07:18:05.946489 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:05Z","lastTransitionTime":"2025-10-02T07:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.050712 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.050777 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.050794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.050818 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.050833 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.080832 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/3.log" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.082289 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/2.log" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.086941 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" exitCode=1 Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.087006 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.087114 4829 scope.go:117] "RemoveContainer" containerID="574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.088301 4829 scope.go:117] "RemoveContainer" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" Oct 02 07:18:06 crc kubenswrapper[4829]: E1002 07:18:06.088556 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.114372 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.145809 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.153030 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.153088 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.153106 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.153129 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.153146 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.163150 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.181003 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.200908 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.217836 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.237295 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.256043 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.256142 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.256161 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.256186 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.256204 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.259138 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.282274 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://574196ed81600f42424b88e79b30c179bd5e414848e114f61f72d19e8f16dd40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:37Z\\\",\\\"message\\\":\\\"6505 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 07:17:37.474898 6505 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 07:17:37.474902 6505 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 07:17:37.474919 6505 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1002 07:17:37.474936 6505 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 07:17:37.474936 6505 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 07:17:37.474953 6505 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 07:17:37.474967 6505 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 07:17:37.475320 6505 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475469 6505 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 07:17:37.475927 6505 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 07:17:37.476360 6505 factory.go:656] Stopping watch factory\\\\nI1002 07:17:37.476383 6505 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:18:05Z\\\",\\\"message\\\":\\\"kube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z]\\\\nI1002 07:18:05.680332 6860 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc after 0 failed attempt(s)\\\\nI1002 07:18:05.680349 6860 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1002 07:18:05.680355 6860 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-mxmxh\\\\nI1002 07:18:05.680366 6860 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-mxmxh in node crc\\\\nI1002 07:18:05.680374 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:18:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.298618 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.316577 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.335881 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.357071 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.359132 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.359167 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.359177 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.359193 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.359207 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.374263 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.395458 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.415098 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.428704 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.442018 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:06Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.462209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.462300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.462316 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.462340 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.462354 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.565019 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.565079 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.565098 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.565123 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.565144 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.668006 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.668072 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.668095 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.668120 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.668142 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.770378 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.770437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.770455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.770480 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.770499 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.872801 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.872865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.872875 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.872890 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.872899 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.975822 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.975857 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.975868 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.975883 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:06 crc kubenswrapper[4829]: I1002 07:18:06.975892 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:06Z","lastTransitionTime":"2025-10-02T07:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.078737 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.078782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.078794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.078812 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.078825 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.093083 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/3.log" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.098211 4829 scope.go:117] "RemoveContainer" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.098493 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.111644 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.127096 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.144133 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.164990 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.186056 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.186675 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.186697 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.186714 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.186729 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.195715 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:18:05Z\\\",\\\"message\\\":\\\"kube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z]\\\\nI1002 07:18:05.680332 6860 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc after 0 failed attempt(s)\\\\nI1002 07:18:05.680349 6860 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1002 07:18:05.680355 6860 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-mxmxh\\\\nI1002 07:18:05.680366 6860 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-mxmxh in node crc\\\\nI1002 07:18:05.680374 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:18:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.218270 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.238704 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.257540 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.270753 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.288162 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.289934 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.289986 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.290004 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.290029 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.290046 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.309471 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.331326 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.349530 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.369767 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.375654 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.375724 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.375749 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.375778 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.375821 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.404582 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.404644 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.409673 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.409729 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.409748 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.409773 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.409790 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.429508 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.435793 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.435856 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.435879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.435909 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.435931 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.437760 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.456819 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.457715 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.459873 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.459917 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.459925 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.460026 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.460265 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.460441 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.460622 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.460790 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.464523 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.464564 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.464580 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.464604 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.464624 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.481010 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.484815 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.489662 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.489702 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.489715 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.489733 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.489745 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.511372 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:07Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:07 crc kubenswrapper[4829]: E1002 07:18:07.511591 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.513399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.513437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.513454 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.513477 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.513493 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.617133 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.617182 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.617198 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.617221 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.617273 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.720436 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.720500 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.720525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.720553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.720571 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.823419 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.823486 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.823509 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.823540 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.823563 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.926867 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.926922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.926944 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.926972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:07 crc kubenswrapper[4829]: I1002 07:18:07.926993 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:07Z","lastTransitionTime":"2025-10-02T07:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.029968 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.030021 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.030038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.030061 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.030078 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.132830 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.132891 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.132908 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.132937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.132956 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.236136 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.236199 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.236250 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.236300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.236322 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.339682 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.339730 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.339746 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.339770 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.339785 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.442278 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.442341 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.442358 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.442743 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.442793 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.546784 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.546834 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.546850 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.546870 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.546888 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.648922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.649040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.649164 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.649657 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.649735 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.753111 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.753186 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.753208 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.753336 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.753378 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.856059 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.856100 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.856112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.856129 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.856141 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.959794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.959848 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.959861 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.959878 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:08 crc kubenswrapper[4829]: I1002 07:18:08.959893 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:08Z","lastTransitionTime":"2025-10-02T07:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.062197 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.062291 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.062312 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.062335 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.062352 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.165042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.165076 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.165085 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.165099 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.165109 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.268534 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.268606 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.268625 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.268656 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.268678 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.372293 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.372380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.372408 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.372446 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.372471 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.460584 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:09 crc kubenswrapper[4829]: E1002 07:18:09.460801 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.460868 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.460982 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:09 crc kubenswrapper[4829]: E1002 07:18:09.461083 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.461123 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:09 crc kubenswrapper[4829]: E1002 07:18:09.461352 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:09 crc kubenswrapper[4829]: E1002 07:18:09.461486 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.474967 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.475025 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.475038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.475059 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.475072 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.482064 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.507404 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:18:05Z\\\",\\\"message\\\":\\\"kube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z]\\\\nI1002 07:18:05.680332 6860 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc after 0 failed attempt(s)\\\\nI1002 07:18:05.680349 6860 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1002 07:18:05.680355 6860 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-mxmxh\\\\nI1002 07:18:05.680366 6860 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-mxmxh in node crc\\\\nI1002 07:18:05.680374 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:18:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.523727 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.539262 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.561524 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.578145 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.578204 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.578222 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.578302 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.578321 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.584913 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.599964 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.618788 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.634927 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.655613 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.675031 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.681202 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.681307 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.681327 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.681351 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.681368 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.691975 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.712639 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.736431 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.752372 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.770731 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.783925 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.783960 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.783968 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.783982 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.783990 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.784773 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.803438 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:09Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.887538 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.887610 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.887627 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.887651 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.887670 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.990501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.990571 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.990593 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.990622 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:09 crc kubenswrapper[4829]: I1002 07:18:09.990645 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:09Z","lastTransitionTime":"2025-10-02T07:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.093673 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.093754 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.093770 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.093792 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.093809 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.195864 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.195901 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.195917 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.195937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.195954 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.298952 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.299000 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.299011 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.299030 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.299042 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.401059 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.401102 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.401111 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.401126 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.401137 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.505275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.505326 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.505344 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.505367 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.505384 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.609782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.609872 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.609897 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.609926 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.609947 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.713292 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.713357 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.713368 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.713385 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.713396 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.816695 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.816784 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.816808 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.816853 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.816877 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.920011 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.920063 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.920073 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.920092 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:10 crc kubenswrapper[4829]: I1002 07:18:10.920106 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:10Z","lastTransitionTime":"2025-10-02T07:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.022854 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.022892 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.022924 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.022942 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.022954 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.125347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.125399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.125411 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.125428 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.125469 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.228089 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.228148 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.228166 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.228195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.228212 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.330931 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.330995 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.331015 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.331038 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.331055 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.434525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.434587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.434609 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.434633 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.434651 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.460382 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.460399 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.460457 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.460530 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:11 crc kubenswrapper[4829]: E1002 07:18:11.460700 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:11 crc kubenswrapper[4829]: E1002 07:18:11.460837 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:11 crc kubenswrapper[4829]: E1002 07:18:11.461054 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:11 crc kubenswrapper[4829]: E1002 07:18:11.461178 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.540432 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.540478 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.540490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.540508 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.540520 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.642007 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.642046 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.642057 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.642071 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.642081 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.744960 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.745311 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.745334 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.745365 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.745388 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.848781 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.848831 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.848848 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.848869 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.848886 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.951803 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.951873 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.951894 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.951963 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:11 crc kubenswrapper[4829]: I1002 07:18:11.951985 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:11Z","lastTransitionTime":"2025-10-02T07:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.054290 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.054328 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.054338 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.054353 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.054363 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.157416 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.157490 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.157505 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.157540 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.157561 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.259984 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.260048 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.260071 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.260101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.260122 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.363163 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.363284 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.363340 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.363364 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.363381 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.465328 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.465361 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.465369 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.465381 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.465389 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.567515 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.567549 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.567557 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.567570 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.567579 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.670519 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.670593 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.670619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.670649 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.670707 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.773034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.773083 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.773095 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.773112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.773124 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.876063 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.876120 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.876130 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.876147 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.876175 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.978668 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.978741 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.978760 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.978785 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:12 crc kubenswrapper[4829]: I1002 07:18:12.978800 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:12Z","lastTransitionTime":"2025-10-02T07:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.081248 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.081302 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.081314 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.081333 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.081346 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.184456 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.184524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.184548 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.184577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.184599 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.261045 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.261099 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261291 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261317 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261334 4829 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261401 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.261385085 +0000 UTC m=+148.601033500 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261420 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261470 4829 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261492 4829 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.261574 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.26154882 +0000 UTC m=+148.601197255 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.287822 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.287860 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.287873 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.287894 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.287907 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.361816 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.362004 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.362039 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.36200255 +0000 UTC m=+148.701650965 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.362098 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.362115 4829 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.362172 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.362154204 +0000 UTC m=+148.701802639 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.362322 4829 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.362412 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.362392511 +0000 UTC m=+148.702040946 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.390515 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.390579 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.390602 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.390631 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.390653 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.459859 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.459899 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.459995 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.460185 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.460276 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.460514 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.460586 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:13 crc kubenswrapper[4829]: E1002 07:18:13.460891 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.474480 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.493484 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.493536 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.493552 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.493575 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.493594 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.596747 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.596804 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.596815 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.596833 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.596848 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.699662 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.699789 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.699818 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.699848 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.699870 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.802879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.802935 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.802946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.802963 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.802975 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.905921 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.906004 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.906017 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.906047 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:13 crc kubenswrapper[4829]: I1002 07:18:13.906063 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:13Z","lastTransitionTime":"2025-10-02T07:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.009422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.009484 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.009501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.009525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.009543 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.112345 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.112405 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.112461 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.112486 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.112508 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.215477 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.215529 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.215545 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.215567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.215585 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.319331 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.319445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.319475 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.319503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.319524 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.422584 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.422654 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.422672 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.422694 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.422711 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.526028 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.526064 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.526076 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.526093 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.526106 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.629503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.629561 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.629578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.629605 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.629631 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.733413 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.733468 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.733487 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.733513 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.733532 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.836867 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.836913 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.836930 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.836954 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.836973 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.940906 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.940970 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.940988 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.941010 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:14 crc kubenswrapper[4829]: I1002 07:18:14.941024 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:14Z","lastTransitionTime":"2025-10-02T07:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.044138 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.044191 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.044209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.044263 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.044284 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.147056 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.147114 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.147135 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.147158 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.147175 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.249740 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.249805 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.249822 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.249843 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.249861 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.352988 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.353050 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.353072 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.353101 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.353123 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.456484 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.456615 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.456641 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.456673 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.456694 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.459815 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.459863 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.459944 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:15 crc kubenswrapper[4829]: E1002 07:18:15.460125 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.460162 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:15 crc kubenswrapper[4829]: E1002 07:18:15.460293 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:15 crc kubenswrapper[4829]: E1002 07:18:15.460487 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:15 crc kubenswrapper[4829]: E1002 07:18:15.460567 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.559440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.559508 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.559528 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.559553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.559570 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.662722 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.662765 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.662782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.662804 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.662822 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.765402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.765445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.765458 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.765474 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.765486 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.871938 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.871993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.872012 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.872035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.872051 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.974475 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.974522 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.974534 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.974553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:15 crc kubenswrapper[4829]: I1002 07:18:15.974566 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:15Z","lastTransitionTime":"2025-10-02T07:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.077455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.077491 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.077503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.077520 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.077531 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.180150 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.180189 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.180198 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.180213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.180249 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.282886 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.282927 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.282935 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.282950 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.282961 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.385508 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.385542 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.385553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.385567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.385577 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.488968 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.489034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.489056 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.489088 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.489109 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.592365 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.592426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.592444 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.592470 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.592491 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.695461 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.695510 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.695529 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.695556 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.695575 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.799029 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.799111 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.799134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.799167 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.799189 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.902164 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.902192 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.902201 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.902215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:16 crc kubenswrapper[4829]: I1002 07:18:16.902244 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:16Z","lastTransitionTime":"2025-10-02T07:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.004977 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.005029 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.005042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.005062 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.005074 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.107569 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.107598 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.107607 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.107619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.107629 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.210758 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.210813 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.210825 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.210842 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.210853 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.313919 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.314043 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.314061 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.314083 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.314103 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.417118 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.417163 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.417177 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.417193 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.417206 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.460390 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.460447 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.460538 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.460617 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.460655 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.460821 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.460897 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.461041 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.520918 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.520964 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.520977 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.520993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.521011 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.573169 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.573213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.573250 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.573265 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.573275 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.585285 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.589532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.589562 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.589572 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.589586 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.589598 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.608410 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.612778 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.612851 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.612876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.612904 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.612920 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.630312 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.634731 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.634767 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.634780 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.634797 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.634809 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.654365 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.658950 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.658997 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.659015 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.659043 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.659060 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.678952 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:17Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:17 crc kubenswrapper[4829]: E1002 07:18:17.679306 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.681378 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.681437 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.681455 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.681480 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.681497 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.783985 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.784023 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.784034 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.784051 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.784066 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.886910 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.886989 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.887008 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.887035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.887052 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.990300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.990372 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.990392 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.990418 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:17 crc kubenswrapper[4829]: I1002 07:18:17.990438 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:17Z","lastTransitionTime":"2025-10-02T07:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.093516 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.093595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.093621 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.093655 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.093678 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.197280 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.197336 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.197354 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.197443 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.197474 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.300547 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.300618 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.300641 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.300671 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.300693 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.404035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.404110 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.404125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.404145 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.404160 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.506893 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.506945 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.506956 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.506977 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.506990 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.610100 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.610170 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.610194 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.610219 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.610269 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.713096 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.713178 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.713209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.713275 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.713299 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.816167 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.816283 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.816314 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.816342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.816362 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.919077 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.919168 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.919190 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.919213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:18 crc kubenswrapper[4829]: I1002 07:18:18.919257 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:18Z","lastTransitionTime":"2025-10-02T07:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.021948 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.022091 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.022112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.022140 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.022157 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.124532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.124591 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.124611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.124638 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.124660 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.227626 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.227688 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.227703 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.227725 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.227738 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.330168 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.330267 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.330280 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.330296 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.330309 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.433154 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.433260 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.433294 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.433324 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.433343 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.459840 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.459897 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.459976 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:19 crc kubenswrapper[4829]: E1002 07:18:19.460045 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.460097 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:19 crc kubenswrapper[4829]: E1002 07:18:19.460247 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:19 crc kubenswrapper[4829]: E1002 07:18:19.460451 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:19 crc kubenswrapper[4829]: E1002 07:18:19.460678 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.480684 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.498796 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.528135 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:18:05Z\\\",\\\"message\\\":\\\"kube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z]\\\\nI1002 07:18:05.680332 6860 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc after 0 failed attempt(s)\\\\nI1002 07:18:05.680349 6860 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1002 07:18:05.680355 6860 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-mxmxh\\\\nI1002 07:18:05.680366 6860 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-mxmxh in node crc\\\\nI1002 07:18:05.680374 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:18:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.536269 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.536415 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.536445 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.536474 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.536495 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.544892 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.561761 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.577283 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f098266b-5241-4134-ae73-b876046e0658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9691cc2ca2cc22e3f216ecc263a80476389bf663aa9a157e945a31f120b6474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a11daceac317a87dfef15bdc352d73221a0793f6951cde667b9d92349f9e97cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11daceac317a87dfef15bdc352d73221a0793f6951cde667b9d92349f9e97cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.597280 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.613104 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.634130 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.638737 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.638799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.638824 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.638855 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.638875 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.651991 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.676385 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.695612 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.709749 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.727273 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.742081 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.742173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.742246 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.742272 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.742290 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.757759 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.776423 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.794677 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.810673 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.824783 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:19Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.845208 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.845283 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.845302 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.845332 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.845355 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.946957 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.946988 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.946997 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.947011 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:19 crc kubenswrapper[4829]: I1002 07:18:19.947019 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:19Z","lastTransitionTime":"2025-10-02T07:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.049469 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.049512 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.049525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.049540 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.049551 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.152041 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.152105 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.152125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.152149 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.152165 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.254778 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.254855 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.254882 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.254914 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.254937 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.357483 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.357517 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.357530 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.357547 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.357559 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.460579 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.460652 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.460675 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.460701 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.460723 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.461290 4829 scope.go:117] "RemoveContainer" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" Oct 02 07:18:20 crc kubenswrapper[4829]: E1002 07:18:20.461537 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.563689 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.563745 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.563764 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.563788 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.563806 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.666946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.667008 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.667032 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.667059 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.667080 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.770228 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.770309 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.770326 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.770349 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.770366 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.873525 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.873584 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.873600 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.873623 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.873639 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.977413 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.977511 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.977534 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.977564 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:20 crc kubenswrapper[4829]: I1002 07:18:20.977589 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:20Z","lastTransitionTime":"2025-10-02T07:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.080539 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.080593 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.080611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.080635 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.080653 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.182987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.183061 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.183124 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.183150 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.183168 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.285825 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.285895 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.285914 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.285939 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.285958 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.389487 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.389537 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.389553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.389575 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.389592 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.459876 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.459909 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.459966 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.460041 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:21 crc kubenswrapper[4829]: E1002 07:18:21.460198 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:21 crc kubenswrapper[4829]: E1002 07:18:21.460402 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:21 crc kubenswrapper[4829]: E1002 07:18:21.460554 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:21 crc kubenswrapper[4829]: E1002 07:18:21.460627 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.492680 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.492731 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.492748 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.492772 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.492788 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.595157 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.595212 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.595264 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.595288 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.595305 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.698033 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.698085 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.698102 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.698129 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.698147 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.801594 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.801650 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.801669 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.801694 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.801712 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.904609 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.904668 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.904686 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.904709 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:21 crc kubenswrapper[4829]: I1002 07:18:21.904727 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:21Z","lastTransitionTime":"2025-10-02T07:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.007534 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.007587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.007605 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.007678 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.007705 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.111122 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.111177 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.111195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.111218 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.111288 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.214163 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.214259 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.214278 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.214306 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.214421 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.318333 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.318392 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.318409 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.318433 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.318455 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.425322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.425402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.425428 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.425473 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.425498 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.528786 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.528855 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.528877 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.528904 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.528927 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.632125 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.632176 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.632195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.632217 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.632346 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.735173 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.735289 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.735309 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.735342 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.735363 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.838819 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.838900 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.838919 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.838957 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.838976 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.942088 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.942179 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.942208 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.942274 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:22 crc kubenswrapper[4829]: I1002 07:18:22.942297 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:22Z","lastTransitionTime":"2025-10-02T07:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.047422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.047509 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.047536 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.047572 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.047598 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.151006 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.151093 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.151121 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.151157 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.151181 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.254392 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.254476 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.254503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.254532 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.254551 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.357768 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.357832 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.357849 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.357875 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.357891 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.461683 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:23 crc kubenswrapper[4829]: E1002 07:18:23.461858 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.462405 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:23 crc kubenswrapper[4829]: E1002 07:18:23.462595 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.462645 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:23 crc kubenswrapper[4829]: E1002 07:18:23.462813 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.462841 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.462820 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: E1002 07:18:23.462989 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.463016 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.463037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.463062 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.463082 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.566767 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.566822 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.566840 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.566864 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.566882 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.670273 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.670637 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.670659 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.670689 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.670707 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.774028 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.774112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.774128 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.774152 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.774170 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.877611 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.877665 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.877681 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.877712 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.877730 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.982130 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.982209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.982280 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.982318 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:23 crc kubenswrapper[4829]: I1002 07:18:23.982348 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:23Z","lastTransitionTime":"2025-10-02T07:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.085501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.085567 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.085586 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.085610 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.085628 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.187823 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.187899 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.187919 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.187944 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.187963 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.290946 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.291002 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.291017 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.291037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.291048 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.393996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.394040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.394052 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.394067 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.394077 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.496952 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.497007 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.497025 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.497049 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.497066 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.598619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.598653 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.598664 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.598679 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.598689 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.700865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.700936 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.700959 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.700989 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.701009 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.804658 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.804719 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.804737 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.804765 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.804788 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.907598 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.907645 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.907661 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.907684 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:24 crc kubenswrapper[4829]: I1002 07:18:24.907701 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:24Z","lastTransitionTime":"2025-10-02T07:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.011357 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.011409 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.011426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.011448 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.011465 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.113973 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.114281 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.114389 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.114488 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.114592 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.223511 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.223585 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.223603 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.223627 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.223643 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.325829 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.325883 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.325900 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.325922 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.325938 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.428925 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.429011 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.429040 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.429071 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.429093 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.460045 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.460056 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.460078 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.460291 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:25 crc kubenswrapper[4829]: E1002 07:18:25.460302 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:25 crc kubenswrapper[4829]: E1002 07:18:25.460388 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:25 crc kubenswrapper[4829]: E1002 07:18:25.460520 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:25 crc kubenswrapper[4829]: E1002 07:18:25.460766 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.531699 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.531765 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.531791 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.531818 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.531839 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.635443 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.635521 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.635545 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.635578 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.635600 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.739287 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.739344 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.739363 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.739387 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.739405 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.842157 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.842270 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.842289 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.842315 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.842333 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.945735 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.945787 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.945805 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.945828 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:25 crc kubenswrapper[4829]: I1002 07:18:25.945847 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:25Z","lastTransitionTime":"2025-10-02T07:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.048868 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.048944 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.048966 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.048996 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.049016 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.152488 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.152549 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.152605 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.152630 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.152647 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.255137 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.255211 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.255268 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.255300 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.255321 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.358491 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.358558 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.358577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.358603 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.358626 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.461569 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.461631 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.461650 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.461671 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.461690 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.564530 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.564586 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.564605 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.564629 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.564645 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.667527 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.667580 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.667597 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.667620 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.667638 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.770074 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.770120 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.770137 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.770159 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.770175 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.873016 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.873115 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.873133 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.873155 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.873172 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.976497 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.976568 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.976587 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.976612 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:26 crc kubenswrapper[4829]: I1002 07:18:26.976629 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:26Z","lastTransitionTime":"2025-10-02T07:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.078849 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.079121 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.079277 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.079414 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.079510 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.182645 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.182718 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.182739 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.182769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.182789 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.286405 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.286477 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.286499 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.286527 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.286551 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.389669 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.389714 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.389730 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.389754 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.389772 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.460015 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.460490 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.460487 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.460536 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.460947 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.460615 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.461141 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.460789 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.493280 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.493330 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.493347 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.493368 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.493384 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.596977 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.597360 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.597508 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.597663 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.597790 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.701025 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.701076 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.701094 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.701117 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.701134 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.743903 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.744014 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.744042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.744071 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.744096 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.761931 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.767837 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.767908 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.767929 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.767954 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.767978 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.788907 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.793042 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.793099 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.793112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.793129 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.793145 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.807752 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.812036 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.812127 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.812145 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.812198 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.812217 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.827820 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.831940 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.832136 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.832349 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.832538 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.832880 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.849205 4829 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"66ad52c8-80dd-4f56-a08f-df32feb03baf\\\",\\\"systemUUID\\\":\\\"95b96f34-f875-43c2-b6d4-5079b80b707a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:27Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:27 crc kubenswrapper[4829]: E1002 07:18:27.849796 4829 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.851768 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.851943 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.852069 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.852190 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.852339 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.955277 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.955976 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.956123 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.956274 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:27 crc kubenswrapper[4829]: I1002 07:18:27.956404 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:27Z","lastTransitionTime":"2025-10-02T07:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.059202 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.059286 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.059302 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.059324 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.059339 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.161634 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.161678 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.161694 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.161719 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.161737 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.265459 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.265788 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.265925 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.266090 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.266212 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.368374 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.368737 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.368811 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.368894 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.368962 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.470987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.471018 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.471032 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.471066 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.471077 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.573597 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.573648 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.573659 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.573679 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.573691 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.676585 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.676653 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.676679 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.676711 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.676732 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.779609 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.779685 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.779695 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.779708 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.779717 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.882116 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.882203 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.882217 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.882250 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.882264 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.984857 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.984908 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.984919 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.984938 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:28 crc kubenswrapper[4829]: I1002 07:18:28.984949 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:28Z","lastTransitionTime":"2025-10-02T07:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.088647 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.088703 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.088724 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.088754 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.088776 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.191210 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.191295 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.191313 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.191340 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.191364 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.293195 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.293484 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.293559 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.293647 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.293704 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.396426 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.396483 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.396501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.396527 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.396547 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.459981 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.460066 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:29 crc kubenswrapper[4829]: E1002 07:18:29.460176 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.460246 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:29 crc kubenswrapper[4829]: E1002 07:18:29.460444 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:29 crc kubenswrapper[4829]: E1002 07:18:29.460575 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.461569 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:29 crc kubenswrapper[4829]: E1002 07:18:29.462070 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.480731 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfc4c0b4-5fa7-47d5-90a6-c43036a784fd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e5737e23f9330a36b8b8ecd9f8bd7542771b76c5d90a7dbed69f531d1196ffb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f70f2673abf503dc15c9d80b391d8830284088871edd52e9607141d7d31e918d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://92be508775be519f7118ed8c5c0180f2b05230a9acc9135796bc94a19220b0dd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7414f9edee711ef99b069b286a31c7abd67590486dd4534c73994d7d2a6231bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.499472 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.499543 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.499565 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.499589 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.499608 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.500785 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.528187 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"00c0190d-cefe-408f-88c0-fa849e3811d3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:18:05Z\\\",\\\"message\\\":\\\"kube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:05Z is after 2025-08-24T17:21:41Z]\\\\nI1002 07:18:05.680332 6860 obj_retry.go:386] Retry successful for *v1.Pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc after 0 failed attempt(s)\\\\nI1002 07:18:05.680349 6860 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1002 07:18:05.680355 6860 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/network-metrics-daemon-mxmxh\\\\nI1002 07:18:05.680366 6860 ovn.go:134] Ensuring zone local for Pod openshift-multus/network-metrics-daemon-mxmxh in node crc\\\\nI1002 07:18:05.680374 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:18:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4smrk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-2lvjk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.541979 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tbzrs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f9a0b933-69ad-4084-a133-df1872f73c29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6d3919587b96eff963e43316e91869ead4dd73bd9c7ba50df7cd9d1ea0f900c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4k7pp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tbzrs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.557253 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"675c4639-4b87-404e-8258-0e5bae51d933\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:25Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwdx6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:25Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mxmxh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.573699 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f098266b-5241-4134-ae73-b876046e0658\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9691cc2ca2cc22e3f216ecc263a80476389bf663aa9a157e945a31f120b6474\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a11daceac317a87dfef15bdc352d73221a0793f6951cde667b9d92349f9e97cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a11daceac317a87dfef15bdc352d73221a0793f6951cde667b9d92349f9e97cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.585986 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:29 crc kubenswrapper[4829]: E1002 07:18:29.586155 4829 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:18:29 crc kubenswrapper[4829]: E1002 07:18:29.586268 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs podName:675c4639-4b87-404e-8258-0e5bae51d933 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:33.586249104 +0000 UTC m=+164.925897509 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs") pod "network-metrics-daemon-mxmxh" (UID: "675c4639-4b87-404e-8258-0e5bae51d933") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.586856 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe0ae4c1727aad6d70dcdd5ee44f00ef22eea9fe30d66f6dc984c88f8e7fd2a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.599542 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f16811bb0e635ee24bf8038a8e2e626689aa647713738e2df7be12ea594e76f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.602730 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.602765 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.602777 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.602792 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.602801 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.613058 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-xrncc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8880512f-81d6-4b87-82b7-4ce1a2a13f9a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5e9bddcfe3d5ddbf89654612a0d8d4d946119ed620c776a349b6039fd2c99c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a459a8eaee5c0339bc9123bef72893d0026290507c875d02c31453eb8935310\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27c9886cf46afa080f629629f6f537ad4f71a7890bfd3783ee27b19451877817\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e5b15b4b70b468fc1d67d5022a81ea08d03a453aa4babf6bf12f45dd9a293a20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://072e8f685f425cc17fbf7268b7669b6b3ebb18d01e43eb5aff168bf5ee2a9e20\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb488d0a45a89deeaecdf3cd60d6b325a3a4772ff2f5e5c62ca40ec678e59a52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff30616335cb976d3cf018d76189f0eb7ed96e9e86fa65dbf42c009701ec66e3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:17:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zsfzw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-xrncc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.626737 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a72492ff-5348-4673-a508-6f55961d534f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0b8e01abc4b9f55f5b98f352d14bc196b7b352d6669de990416c302b830dfd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebcc71c35ccd6fa26254a170c3237840edd9d8babeee87be7e485b97dc78e969\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ab9be0018d67cf27279c9a8f56adabce12292e7dd970da6193bc6a08de71954\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b3d748eb40a51fed393178a74527c4980fa72bfad725d4789c41c6dba3b7c8a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee1c1216b26ab8f316e293be7164ff34b3b5980a18ae1d74c6e978c94b9b6621\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"file observer\\\\nW1002 07:17:09.358915 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1002 07:17:09.359085 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 07:17:09.359919 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-202610342/tls.crt::/tmp/serving-cert-202610342/tls.key\\\\\\\"\\\\nI1002 07:17:09.594075 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 07:17:09.598748 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 07:17:09.598779 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 07:17:09.598800 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 07:17:09.598806 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 07:17:09.606751 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 07:17:09.606890 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606918 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 07:17:09.606940 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1002 07:17:09.606959 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 07:17:09.606980 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 07:17:09.607000 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1002 07:17:09.606849 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1002 07:17:09.612413 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:03Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1400fdd28461517de6aa93509a57af595e67c11045fbfa6244a3349b71de64\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105b550532b4befcf1081b260d694475c67e8768939d881597dc21ab90335c74\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.639433 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.653518 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.662645 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vshhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87f69905-b2b5-4012-9b1d-77e8d22b2cb3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1284a4ab4c534ba570969b22004743f44a72f27ca312329d2970bfab7e787c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7qt7q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vshhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.672821 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"60aa0a81-354e-46f1-ab0c-b1eb386974a6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6fec140afc82e7ed34c1e4dd1baf7dd85e2f090e82c3d317e0453e4347028d09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9jlp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8rvq4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.691168 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae3a9066-e6ae-42fc-b1da-2d3342ccff1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bdfb51f3f6bf31537cb0a9785510d82887fe483d499e9846a23e6c552c3bffab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1fa816206b5d4d3d505205963ff22cea1671bd694a5d7b14b4685820c8c6751b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a1ba2a401204816dd48371bd60b60c56ebd12de96ef9c3f532df865b52e69d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5042b51c03224550f4977031fd2e1efe8610ff29ea375c7170dc088ecf1c3eb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81dbf5d9cffa3a23974ba43976a94bfbfe80a1b6d42494a20638a7a5dfabd81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e28e8d9250d3d2c1a42e0a61dc761aa48934b7b12df547b9c77a7ac5fb865653\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8de544c201d7eb3e9696f418582ec099030a86c8d00fdb5c1ee9f4b74fc2d8b3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b209a95f25cd71a4ef279eb9e53647bf5835be697d0bb3b4a4c1328a7356fffa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:52Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.705304 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.705354 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.705371 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.705395 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.705411 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.706065 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18e5c3ec-9c80-4123-8f12-413920595abd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:16:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0bd54effe4924fde8a4baa1cb4bea54432e7d119bc010f44c62baaf4a2a0c5a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd99ca93b783ceca34ed45b9e4927ef21839e25b5aa250693aebb603f03bdb8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e49acbcc7b5d6845f21be5162be1b451d7998cf50b9ac780cb0a62dfbe1aff06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:16:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b270faf0db2204254de5d809a914b29b1f4dc2d6ba549727e908782646524a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T07:16:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T07:16:50Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:16:49Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.718451 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:10Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be6055f783d93f91094592caf023d587b31b1e00d52ae39e9fc8991d942fde25\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ebf7eb85adaa8ab61579132640561f2061eaaa50a35c3a5f7b041a75a98a05d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.729667 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-spzjt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ca00dadf-1664-466b-830c-e172857db47a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T07:17:58Z\\\",\\\"message\\\":\\\"2025-10-02T07:17:12+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818\\\\n2025-10-02T07:17:12+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_36e4dd46-1b47-4aa4-a594-808be0aab818 to /host/opt/cni/bin/\\\\n2025-10-02T07:17:13Z [verbose] multus-daemon started\\\\n2025-10-02T07:17:13Z [verbose] Readiness Indicator file check\\\\n2025-10-02T07:17:58Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T07:17:11Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-whzpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-spzjt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.740194 4829 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c45bd15-86f4-4d9c-8dfc-04274acd600f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T07:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://053ae3e25f7dc283db86c32b7f3f798e4f74f4949d3ea7a996632a11d9142fc2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0fb5b4b93a65863b755f49cd4025ab3db6a26d07954bc7d89fda8abe7057809\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T07:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-24gn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T07:17:23Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-2b4sc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T07:18:29Z is after 2025-08-24T17:21:41Z" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.807615 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.807655 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.807668 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.807684 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.807696 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.910281 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.910331 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.910348 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.910367 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:29 crc kubenswrapper[4829]: I1002 07:18:29.910382 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:29Z","lastTransitionTime":"2025-10-02T07:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.013343 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.013421 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.013440 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.013466 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.013486 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.115841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.115899 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.115915 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.115939 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.115954 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.218686 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.218752 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.218777 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.218806 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.218829 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.322062 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.322123 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.322140 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.322165 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.322185 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.424564 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.424837 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.425043 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.425313 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.425573 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.528596 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.528876 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.528972 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.529065 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.529152 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.631994 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.632055 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.632075 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.632100 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.632119 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.735292 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.735332 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.735340 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.735355 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.735365 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.837991 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.838077 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.838097 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.838119 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.838137 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.941118 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.941243 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.941265 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.941287 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:30 crc kubenswrapper[4829]: I1002 07:18:30.941303 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:30Z","lastTransitionTime":"2025-10-02T07:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.044278 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.044373 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.044400 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.044433 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.044456 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.147582 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.147649 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.147672 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.147701 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.147721 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.250359 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.250425 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.250449 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.250470 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.250486 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.354742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.354834 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.354854 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.354878 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.354896 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.457858 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.457935 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.457960 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.457990 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.458012 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.460721 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.460770 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:31 crc kubenswrapper[4829]: E1002 07:18:31.460898 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.460979 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:31 crc kubenswrapper[4829]: E1002 07:18:31.461087 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.461272 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:31 crc kubenswrapper[4829]: E1002 07:18:31.461360 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:31 crc kubenswrapper[4829]: E1002 07:18:31.461529 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.561416 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.561489 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.561507 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.561530 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.561546 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.665010 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.665067 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.665093 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.665164 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.665190 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.768012 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.768065 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.768086 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.768110 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.768128 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.870654 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.870718 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.870742 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.870773 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.870795 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.973737 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.973814 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.973842 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.973873 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:31 crc kubenswrapper[4829]: I1002 07:18:31.973894 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:31Z","lastTransitionTime":"2025-10-02T07:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.076581 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.076634 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.076653 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.076676 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.076696 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.179631 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.179697 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.179724 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.179755 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.179776 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.282691 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.282751 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.282770 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.282794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.282813 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.385424 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.385471 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.385482 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.385499 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.385512 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.487951 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.488006 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.488019 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.488037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.488049 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.591255 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.591303 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.591317 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.591333 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.591345 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.700035 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.700083 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.700095 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.700113 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.700124 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.802295 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.802380 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.802399 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.802422 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.802440 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.904833 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.904881 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.904893 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.904910 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:32 crc kubenswrapper[4829]: I1002 07:18:32.904925 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:32Z","lastTransitionTime":"2025-10-02T07:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.007822 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.007879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.007898 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.007964 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.008000 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.110654 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.110701 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.110712 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.110731 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.110743 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.213624 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.213664 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.213675 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.213690 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.213701 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.316732 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.316782 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.316801 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.316823 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.316851 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.420117 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.420186 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.420209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.420276 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.420301 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.460203 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.460294 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.460215 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:33 crc kubenswrapper[4829]: E1002 07:18:33.460422 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.460476 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:33 crc kubenswrapper[4829]: E1002 07:18:33.460611 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:33 crc kubenswrapper[4829]: E1002 07:18:33.460832 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:33 crc kubenswrapper[4829]: E1002 07:18:33.460915 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.522751 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.522794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.522806 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.522824 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.522836 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.625127 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.625199 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.625215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.625269 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.625282 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.728037 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.728093 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.728112 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.728136 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.728153 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.831197 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.831369 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.831392 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.831420 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.831438 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.934544 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.934595 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.934612 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.934634 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:33 crc kubenswrapper[4829]: I1002 07:18:33.934654 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:33Z","lastTransitionTime":"2025-10-02T07:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.071439 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.071501 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.071524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.071553 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.071580 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.173756 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.173794 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.173807 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.173821 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.173829 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.276593 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.276637 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.276646 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.276662 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.276671 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.380073 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.380119 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.380130 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.380146 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.380160 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.483161 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.483202 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.483215 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.483246 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.483258 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.585140 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.585193 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.585211 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.585259 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.585277 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.688154 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.688213 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.688260 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.688285 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.688302 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.790637 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.790684 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.790698 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.790714 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.790726 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.893432 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.893480 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.893495 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.893517 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.893535 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.996305 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.996370 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.996391 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.996418 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:34 crc kubenswrapper[4829]: I1002 07:18:34.996441 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:34Z","lastTransitionTime":"2025-10-02T07:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.098865 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.098936 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.098959 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.098987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.099008 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.201626 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.201672 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.201688 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.201710 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.201727 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.304095 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.304142 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.304156 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.304176 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.304191 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.407204 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.407250 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.407259 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.407271 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.407279 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.460384 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.460414 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:35 crc kubenswrapper[4829]: E1002 07:18:35.460582 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.460679 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:35 crc kubenswrapper[4829]: E1002 07:18:35.460853 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.461105 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:35 crc kubenswrapper[4829]: E1002 07:18:35.461509 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:35 crc kubenswrapper[4829]: E1002 07:18:35.461685 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.462137 4829 scope.go:117] "RemoveContainer" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" Oct 02 07:18:35 crc kubenswrapper[4829]: E1002 07:18:35.462471 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-2lvjk_openshift-ovn-kubernetes(00c0190d-cefe-408f-88c0-fa849e3811d3)\"" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.510009 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.510076 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.510102 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.510134 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.510157 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.613574 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.613796 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.613807 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.613820 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.613829 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.716469 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.716519 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.716530 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.716548 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.716560 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.819075 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.819132 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.819145 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.819165 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.819177 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.921392 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.921447 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.921463 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.921486 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:35 crc kubenswrapper[4829]: I1002 07:18:35.921502 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:35Z","lastTransitionTime":"2025-10-02T07:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.024203 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.024273 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.024289 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.024322 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.024335 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.127159 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.127199 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.127209 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.127248 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.127265 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.229889 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.229948 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.229969 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.229999 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.230074 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.333665 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.333727 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.333743 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.333769 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.333785 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.437515 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.437574 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.437591 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.437615 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.437667 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.540824 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.540879 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.540897 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.540923 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.540968 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.644518 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.644673 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.644702 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.644901 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.645066 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.748584 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.748657 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.748669 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.748685 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.748696 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.852503 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.852546 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.852577 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.852596 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.852607 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.955149 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.955254 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.955277 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.955330 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:36 crc kubenswrapper[4829]: I1002 07:18:36.955348 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:36Z","lastTransitionTime":"2025-10-02T07:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.058993 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.059048 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.059062 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.059081 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.059095 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.162170 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.162267 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.162290 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.162317 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.162340 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.265783 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.265841 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.265870 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.265918 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.265940 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.368319 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.368370 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.368383 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.368402 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.368415 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.460074 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.460097 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.460281 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:37 crc kubenswrapper[4829]: E1002 07:18:37.460353 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.460374 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:37 crc kubenswrapper[4829]: E1002 07:18:37.460525 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:37 crc kubenswrapper[4829]: E1002 07:18:37.460610 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:37 crc kubenswrapper[4829]: E1002 07:18:37.460686 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.474688 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.474749 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.474777 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.474799 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.474815 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.578405 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.578469 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.578487 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.578514 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.578531 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.681990 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.682146 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.682166 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.682191 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.682256 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.785987 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.786062 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.786085 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.786116 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.786145 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.889686 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.889775 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.889795 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.889848 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.889867 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.993806 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.993912 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.993937 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.993962 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:37 crc kubenswrapper[4829]: I1002 07:18:37.993979 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:37Z","lastTransitionTime":"2025-10-02T07:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.097591 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.097677 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.097695 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.097750 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.097768 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:38Z","lastTransitionTime":"2025-10-02T07:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.201582 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.201638 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.201660 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.201683 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.201700 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:38Z","lastTransitionTime":"2025-10-02T07:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.209412 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.209477 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.209498 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.209524 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.209542 4829 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T07:18:38Z","lastTransitionTime":"2025-10-02T07:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.276269 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh"] Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.278203 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.280484 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.281082 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.282071 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.284037 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.337662 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.33763404 podStartE2EDuration="1m28.33763404s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.315002709 +0000 UTC m=+109.654651164" watchObservedRunningTime="2025-10-02 07:18:38.33763404 +0000 UTC m=+109.677282485" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.390277 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-vshhb" podStartSLOduration=89.390254914 podStartE2EDuration="1m29.390254914s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.375666979 +0000 UTC m=+109.715315424" watchObservedRunningTime="2025-10-02 07:18:38.390254914 +0000 UTC m=+109.729903319" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.420477 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/29c04574-d195-4a5c-8a79-d77c4879ac64-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.420562 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/29c04574-d195-4a5c-8a79-d77c4879ac64-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.420610 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/29c04574-d195-4a5c-8a79-d77c4879ac64-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.420696 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29c04574-d195-4a5c-8a79-d77c4879ac64-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.420737 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/29c04574-d195-4a5c-8a79-d77c4879ac64-service-ca\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.434450 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podStartSLOduration=89.43443280299999 podStartE2EDuration="1m29.434432803s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.390205823 +0000 UTC m=+109.729854258" watchObservedRunningTime="2025-10-02 07:18:38.434432803 +0000 UTC m=+109.774081218" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.434619 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=89.434612529 podStartE2EDuration="1m29.434612529s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.434011551 +0000 UTC m=+109.773660006" watchObservedRunningTime="2025-10-02 07:18:38.434612529 +0000 UTC m=+109.774260954" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.468582 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=54.468565269 podStartE2EDuration="54.468565269s" podCreationTimestamp="2025-10-02 07:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.452848411 +0000 UTC m=+109.792496826" watchObservedRunningTime="2025-10-02 07:18:38.468565269 +0000 UTC m=+109.808213694" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.485265 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-spzjt" podStartSLOduration=89.485247005 podStartE2EDuration="1m29.485247005s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.484892975 +0000 UTC m=+109.824541450" watchObservedRunningTime="2025-10-02 07:18:38.485247005 +0000 UTC m=+109.824895420" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.502212 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-2b4sc" podStartSLOduration=89.502180239 podStartE2EDuration="1m29.502180239s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.501382296 +0000 UTC m=+109.841030741" watchObservedRunningTime="2025-10-02 07:18:38.502180239 +0000 UTC m=+109.841828684" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521319 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/29c04574-d195-4a5c-8a79-d77c4879ac64-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521427 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/29c04574-d195-4a5c-8a79-d77c4879ac64-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521484 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/29c04574-d195-4a5c-8a79-d77c4879ac64-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521498 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29c04574-d195-4a5c-8a79-d77c4879ac64-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521582 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/29c04574-d195-4a5c-8a79-d77c4879ac64-service-ca\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521686 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/29c04574-d195-4a5c-8a79-d77c4879ac64-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.521757 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/29c04574-d195-4a5c-8a79-d77c4879ac64-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.523156 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/29c04574-d195-4a5c-8a79-d77c4879ac64-service-ca\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.542751 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/29c04574-d195-4a5c-8a79-d77c4879ac64-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.544460 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29c04574-d195-4a5c-8a79-d77c4879ac64-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-86fbh\" (UID: \"29c04574-d195-4a5c-8a79-d77c4879ac64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.567040 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=88.567017171 podStartE2EDuration="1m28.567017171s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.549540111 +0000 UTC m=+109.889188556" watchObservedRunningTime="2025-10-02 07:18:38.567017171 +0000 UTC m=+109.906665586" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.605573 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tbzrs" podStartSLOduration=89.605550454 podStartE2EDuration="1m29.605550454s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.604535085 +0000 UTC m=+109.944183490" watchObservedRunningTime="2025-10-02 07:18:38.605550454 +0000 UTC m=+109.945198859" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.606102 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.647059 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=25.647030454 podStartE2EDuration="25.647030454s" podCreationTimestamp="2025-10-02 07:18:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.646151869 +0000 UTC m=+109.985800284" watchObservedRunningTime="2025-10-02 07:18:38.647030454 +0000 UTC m=+109.986678869" Oct 02 07:18:38 crc kubenswrapper[4829]: I1002 07:18:38.695037 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-xrncc" podStartSLOduration=89.695012375 podStartE2EDuration="1m29.695012375s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:38.694698645 +0000 UTC m=+110.034347100" watchObservedRunningTime="2025-10-02 07:18:38.695012375 +0000 UTC m=+110.034660800" Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.208909 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" event={"ID":"29c04574-d195-4a5c-8a79-d77c4879ac64","Type":"ContainerStarted","Data":"09432daeaf7cf48fcc6a46700a4d11d88939073da0d8b6898335747ee7d3fba8"} Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.208966 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" event={"ID":"29c04574-d195-4a5c-8a79-d77c4879ac64","Type":"ContainerStarted","Data":"0c063879c870f892c7b72b6ee9613956b71380ec2b255dd18bf3bec8c5b265d2"} Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.228861 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-86fbh" podStartSLOduration=90.228824795 podStartE2EDuration="1m30.228824795s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:39.228723193 +0000 UTC m=+110.568371638" watchObservedRunningTime="2025-10-02 07:18:39.228824795 +0000 UTC m=+110.568473220" Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.459865 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.459882 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.459952 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:39 crc kubenswrapper[4829]: I1002 07:18:39.459968 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:39 crc kubenswrapper[4829]: E1002 07:18:39.461550 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:39 crc kubenswrapper[4829]: E1002 07:18:39.461848 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:39 crc kubenswrapper[4829]: E1002 07:18:39.461970 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:39 crc kubenswrapper[4829]: E1002 07:18:39.462069 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:41 crc kubenswrapper[4829]: I1002 07:18:41.459948 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:41 crc kubenswrapper[4829]: E1002 07:18:41.460408 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:41 crc kubenswrapper[4829]: I1002 07:18:41.460146 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:41 crc kubenswrapper[4829]: I1002 07:18:41.460178 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:41 crc kubenswrapper[4829]: E1002 07:18:41.460668 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:41 crc kubenswrapper[4829]: E1002 07:18:41.460780 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:41 crc kubenswrapper[4829]: I1002 07:18:41.460155 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:41 crc kubenswrapper[4829]: E1002 07:18:41.460963 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:43 crc kubenswrapper[4829]: I1002 07:18:43.460453 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:43 crc kubenswrapper[4829]: I1002 07:18:43.460541 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:43 crc kubenswrapper[4829]: I1002 07:18:43.461445 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:43 crc kubenswrapper[4829]: E1002 07:18:43.461709 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:43 crc kubenswrapper[4829]: I1002 07:18:43.461761 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:43 crc kubenswrapper[4829]: E1002 07:18:43.461945 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:43 crc kubenswrapper[4829]: E1002 07:18:43.462099 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:43 crc kubenswrapper[4829]: E1002 07:18:43.462204 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.231008 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/1.log" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.232560 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/0.log" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.232633 4829 generic.go:334] "Generic (PLEG): container finished" podID="ca00dadf-1664-466b-830c-e172857db47a" containerID="3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c" exitCode=1 Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.232680 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerDied","Data":"3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c"} Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.232728 4829 scope.go:117] "RemoveContainer" containerID="983859a69e3cdc923e5f0fee68424b78be3ff947d581be51e98a97da3c44f428" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.233324 4829 scope.go:117] "RemoveContainer" containerID="3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c" Oct 02 07:18:45 crc kubenswrapper[4829]: E1002 07:18:45.233568 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-spzjt_openshift-multus(ca00dadf-1664-466b-830c-e172857db47a)\"" pod="openshift-multus/multus-spzjt" podUID="ca00dadf-1664-466b-830c-e172857db47a" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.459950 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.460047 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:45 crc kubenswrapper[4829]: E1002 07:18:45.460160 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.460179 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:45 crc kubenswrapper[4829]: I1002 07:18:45.460315 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:45 crc kubenswrapper[4829]: E1002 07:18:45.460422 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:45 crc kubenswrapper[4829]: E1002 07:18:45.460525 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:45 crc kubenswrapper[4829]: E1002 07:18:45.460723 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:46 crc kubenswrapper[4829]: I1002 07:18:46.238189 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/1.log" Oct 02 07:18:47 crc kubenswrapper[4829]: I1002 07:18:47.460817 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:47 crc kubenswrapper[4829]: I1002 07:18:47.460848 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:47 crc kubenswrapper[4829]: I1002 07:18:47.460839 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:47 crc kubenswrapper[4829]: E1002 07:18:47.461049 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:47 crc kubenswrapper[4829]: E1002 07:18:47.461158 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:47 crc kubenswrapper[4829]: I1002 07:18:47.461187 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:47 crc kubenswrapper[4829]: E1002 07:18:47.461312 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:47 crc kubenswrapper[4829]: E1002 07:18:47.461442 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:49 crc kubenswrapper[4829]: I1002 07:18:49.461567 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:49 crc kubenswrapper[4829]: I1002 07:18:49.461613 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:49 crc kubenswrapper[4829]: I1002 07:18:49.461666 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:49 crc kubenswrapper[4829]: I1002 07:18:49.461686 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:49 crc kubenswrapper[4829]: E1002 07:18:49.465102 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:49 crc kubenswrapper[4829]: E1002 07:18:49.465286 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:49 crc kubenswrapper[4829]: E1002 07:18:49.465414 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:49 crc kubenswrapper[4829]: E1002 07:18:49.465543 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:49 crc kubenswrapper[4829]: E1002 07:18:49.484497 4829 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 07:18:49 crc kubenswrapper[4829]: E1002 07:18:49.554873 4829 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:18:50 crc kubenswrapper[4829]: I1002 07:18:50.462108 4829 scope.go:117] "RemoveContainer" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.261105 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/3.log" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.265422 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerStarted","Data":"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d"} Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.266292 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.307302 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podStartSLOduration=102.307278621 podStartE2EDuration="1m42.307278621s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:18:51.30586552 +0000 UTC m=+122.645513935" watchObservedRunningTime="2025-10-02 07:18:51.307278621 +0000 UTC m=+122.646927046" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.460539 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.460588 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.460645 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:51 crc kubenswrapper[4829]: E1002 07:18:51.460670 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.460767 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:51 crc kubenswrapper[4829]: E1002 07:18:51.460847 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:51 crc kubenswrapper[4829]: E1002 07:18:51.460973 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:51 crc kubenswrapper[4829]: E1002 07:18:51.461177 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:51 crc kubenswrapper[4829]: I1002 07:18:51.528324 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mxmxh"] Oct 02 07:18:52 crc kubenswrapper[4829]: I1002 07:18:52.269050 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:52 crc kubenswrapper[4829]: E1002 07:18:52.269706 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:53 crc kubenswrapper[4829]: I1002 07:18:53.459918 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:53 crc kubenswrapper[4829]: E1002 07:18:53.460126 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:53 crc kubenswrapper[4829]: I1002 07:18:53.460478 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:53 crc kubenswrapper[4829]: E1002 07:18:53.460576 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:53 crc kubenswrapper[4829]: I1002 07:18:53.460769 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:53 crc kubenswrapper[4829]: E1002 07:18:53.460850 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:54 crc kubenswrapper[4829]: I1002 07:18:54.460163 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:54 crc kubenswrapper[4829]: E1002 07:18:54.460624 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:54 crc kubenswrapper[4829]: E1002 07:18:54.556401 4829 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:18:55 crc kubenswrapper[4829]: I1002 07:18:55.460180 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:55 crc kubenswrapper[4829]: I1002 07:18:55.460241 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:55 crc kubenswrapper[4829]: E1002 07:18:55.460348 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:55 crc kubenswrapper[4829]: I1002 07:18:55.460419 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:55 crc kubenswrapper[4829]: E1002 07:18:55.460479 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:55 crc kubenswrapper[4829]: E1002 07:18:55.460566 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:56 crc kubenswrapper[4829]: I1002 07:18:56.459873 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:56 crc kubenswrapper[4829]: E1002 07:18:56.460333 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:57 crc kubenswrapper[4829]: I1002 07:18:57.460913 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:57 crc kubenswrapper[4829]: I1002 07:18:57.460993 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:57 crc kubenswrapper[4829]: E1002 07:18:57.461053 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:57 crc kubenswrapper[4829]: E1002 07:18:57.461119 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:57 crc kubenswrapper[4829]: I1002 07:18:57.460914 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:57 crc kubenswrapper[4829]: E1002 07:18:57.461262 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:58 crc kubenswrapper[4829]: I1002 07:18:58.460055 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:18:58 crc kubenswrapper[4829]: E1002 07:18:58.460416 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:18:58 crc kubenswrapper[4829]: I1002 07:18:58.460733 4829 scope.go:117] "RemoveContainer" containerID="3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c" Oct 02 07:18:59 crc kubenswrapper[4829]: I1002 07:18:59.295941 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/1.log" Oct 02 07:18:59 crc kubenswrapper[4829]: I1002 07:18:59.296044 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerStarted","Data":"5521c63b7a6584e5b77145dac98a81205b38da465e7c934ddd3fe3ad92d997be"} Oct 02 07:18:59 crc kubenswrapper[4829]: I1002 07:18:59.460523 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:18:59 crc kubenswrapper[4829]: I1002 07:18:59.460567 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:18:59 crc kubenswrapper[4829]: E1002 07:18:59.461786 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:18:59 crc kubenswrapper[4829]: I1002 07:18:59.461881 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:18:59 crc kubenswrapper[4829]: E1002 07:18:59.461940 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:18:59 crc kubenswrapper[4829]: E1002 07:18:59.462065 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:18:59 crc kubenswrapper[4829]: E1002 07:18:59.557114 4829 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:19:00 crc kubenswrapper[4829]: I1002 07:19:00.460141 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:00 crc kubenswrapper[4829]: E1002 07:19:00.460357 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:19:01 crc kubenswrapper[4829]: I1002 07:19:01.460422 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:01 crc kubenswrapper[4829]: I1002 07:19:01.460464 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:01 crc kubenswrapper[4829]: I1002 07:19:01.460594 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:19:01 crc kubenswrapper[4829]: E1002 07:19:01.460815 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:19:01 crc kubenswrapper[4829]: E1002 07:19:01.460934 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:19:01 crc kubenswrapper[4829]: E1002 07:19:01.461074 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:19:02 crc kubenswrapper[4829]: I1002 07:19:02.460406 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:02 crc kubenswrapper[4829]: E1002 07:19:02.460883 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:19:03 crc kubenswrapper[4829]: I1002 07:19:03.460860 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:19:03 crc kubenswrapper[4829]: I1002 07:19:03.460914 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:03 crc kubenswrapper[4829]: I1002 07:19:03.460865 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:03 crc kubenswrapper[4829]: E1002 07:19:03.461034 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 07:19:03 crc kubenswrapper[4829]: E1002 07:19:03.461131 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 07:19:03 crc kubenswrapper[4829]: E1002 07:19:03.461188 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 07:19:04 crc kubenswrapper[4829]: I1002 07:19:04.460351 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:04 crc kubenswrapper[4829]: E1002 07:19:04.460655 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mxmxh" podUID="675c4639-4b87-404e-8258-0e5bae51d933" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.459964 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.460085 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.459993 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.463020 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.463509 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.463744 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 07:19:05 crc kubenswrapper[4829]: I1002 07:19:05.464046 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 07:19:06 crc kubenswrapper[4829]: I1002 07:19:06.460533 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:06 crc kubenswrapper[4829]: I1002 07:19:06.463445 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 07:19:06 crc kubenswrapper[4829]: I1002 07:19:06.463659 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.614619 4829 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.665747 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-l2x2j"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.666574 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.683128 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hwmtz"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.684056 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: W1002 07:19:08.687559 4829 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 02 07:19:08 crc kubenswrapper[4829]: E1002 07:19:08.687660 4829 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:19:08 crc kubenswrapper[4829]: W1002 07:19:08.690697 4829 reflector.go:561] object-"openshift-apiserver"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 02 07:19:08 crc kubenswrapper[4829]: E1002 07:19:08.690782 4829 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.691403 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.691629 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.692633 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.692849 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.693170 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.693668 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.693870 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.694218 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.691476 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.692707 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.693986 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.694884 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.695322 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.695986 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.696079 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.698608 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.698920 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.699323 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.699619 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.703221 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.706862 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l55b8"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.706964 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.707902 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qx6xb"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.708297 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.708595 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.708782 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.709114 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.710962 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2tvsb"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.711461 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.713198 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.713719 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.715322 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.715493 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.715745 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.716028 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.716352 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.716501 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.716023 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.716724 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.716936 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.717557 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.717735 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.717801 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.717747 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.724923 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.726299 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.726733 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.726850 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.726978 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.726741 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.727845 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rnq88"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728181 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728173 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728692 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728289 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728324 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728823 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728409 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728456 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.728500 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.729414 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-226j2"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.729526 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.729679 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.729906 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730172 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730201 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-h2g87"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730309 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730519 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730633 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730751 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.730897 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731040 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731115 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731170 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731377 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731376 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731511 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731707 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.731720 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.732047 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsrrk"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.732374 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.749025 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.759027 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cmbx8"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.759750 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.761374 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.768677 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.774623 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.775051 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.775254 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.775867 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.776385 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.776653 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.776987 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.777323 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.778312 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.778771 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-r7cxf"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.779137 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.779212 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.779818 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.779831 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.779957 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.780072 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.780260 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.782672 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-26l22"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.784572 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.784840 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.785219 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.785562 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.785679 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786027 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786151 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786267 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/20cb3658-2a8d-49c1-8e8e-a7675d34833c-machine-approver-tls\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786304 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/34bc2780-c7b8-45cf-9924-4b08def7fd64-available-featuregates\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786331 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786351 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5443f78e-7898-4ee6-9f8f-574331118acd-audit-dir\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786355 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786508 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786616 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786653 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786766 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786846 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786906 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786961 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786985 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786999 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.786375 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-service-ca\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787055 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-trusted-ca-bundle\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787067 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787084 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5443f78e-7898-4ee6-9f8f-574331118acd-node-pullsecrets\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787103 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787108 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-config\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787128 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6057de60-fc56-49b5-8843-a9e838186747-config\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787139 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787147 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-config\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787150 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787204 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-client\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787240 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787276 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6057de60-fc56-49b5-8843-a9e838186747-images\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787295 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-oauth-serving-cert\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787316 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh87p\" (UniqueName: \"kubernetes.io/projected/ed93e809-cf53-450c-8993-5a69e97542bd-kube-api-access-xh87p\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787350 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v85pb\" (UniqueName: \"kubernetes.io/projected/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-kube-api-access-v85pb\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787367 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-495kt\" (UniqueName: \"kubernetes.io/projected/a5d04969-439e-47d2-8583-875a3f0ae3d5-kube-api-access-495kt\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787727 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-serving-ca\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787776 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787780 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787835 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787884 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787907 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr2g8\" (UniqueName: \"kubernetes.io/projected/34bc2780-c7b8-45cf-9924-4b08def7fd64-kube-api-access-xr2g8\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787925 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787944 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6057de60-fc56-49b5-8843-a9e838186747-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787960 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.787976 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788005 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-serving-cert\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788031 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-image-import-ca\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788009 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788080 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788080 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788017 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788108 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-client-ca\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788166 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-config\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788195 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-encryption-config\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788272 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2njm\" (UniqueName: \"kubernetes.io/projected/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-kube-api-access-l2njm\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788297 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-serving-cert\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788316 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e908e8d-4252-422b-9d40-f38ef0766dde-serving-cert\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788333 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cb3658-2a8d-49c1-8e8e-a7675d34833c-config\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788353 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-oauth-config\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788372 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jlf2\" (UniqueName: \"kubernetes.io/projected/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-kube-api-access-4jlf2\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788397 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/900dfada-c31e-41ec-97a4-2244cf0d8781-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-49p7f\" (UID: \"900dfada-c31e-41ec-97a4-2244cf0d8781\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788413 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s8tg\" (UniqueName: \"kubernetes.io/projected/6e908e8d-4252-422b-9d40-f38ef0766dde-kube-api-access-4s8tg\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788429 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/20cb3658-2a8d-49c1-8e8e-a7675d34833c-auth-proxy-config\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788445 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-serving-cert\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788461 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftw5q\" (UniqueName: \"kubernetes.io/projected/900dfada-c31e-41ec-97a4-2244cf0d8781-kube-api-access-ftw5q\") pod \"cluster-samples-operator-665b6dd947-49p7f\" (UID: \"900dfada-c31e-41ec-97a4-2244cf0d8781\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788477 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-config\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788495 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/96e0038e-e6bd-497c-8245-8f11f8428a00-audit-dir\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788509 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788316 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788490 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788528 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mxc4\" (UniqueName: \"kubernetes.io/projected/5443f78e-7898-4ee6-9f8f-574331118acd-kube-api-access-2mxc4\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788674 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788702 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-audit-policies\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788726 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-etcd-client\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788752 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-config\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788774 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788802 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788828 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed93e809-cf53-450c-8993-5a69e97542bd-serving-cert\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788859 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-ca\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788878 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5d04969-439e-47d2-8583-875a3f0ae3d5-serving-cert\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788906 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gr87\" (UniqueName: \"kubernetes.io/projected/6057de60-fc56-49b5-8843-a9e838186747-kube-api-access-2gr87\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788928 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-service-ca-bundle\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788952 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-encryption-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788976 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-audit\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.788997 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34bc2780-c7b8-45cf-9924-4b08def7fd64-serving-cert\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.789019 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-serving-cert\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.789040 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-service-ca\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.789066 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6nc6\" (UniqueName: \"kubernetes.io/projected/20cb3658-2a8d-49c1-8e8e-a7675d34833c-kube-api-access-n6nc6\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.789093 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v9c4\" (UniqueName: \"kubernetes.io/projected/96e0038e-e6bd-497c-8245-8f11f8428a00-kube-api-access-6v9c4\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.789114 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-trusted-ca\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.789532 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6w49n"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.790165 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.790187 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.790194 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.791246 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.791310 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.791675 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.791694 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.794000 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.797803 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dk9k6"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.798518 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.799000 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.803978 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.807361 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.807611 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.808825 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.809201 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.809898 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.812372 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.834947 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.837785 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.838205 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.839456 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.840657 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.841024 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.841202 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.841512 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.844936 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.845053 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.849369 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r77tn"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.850007 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.854548 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.856082 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.857705 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.860384 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6fsbb"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.860810 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.861176 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.865132 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.865670 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.867204 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.869686 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.870462 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.872749 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsrrk"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.874781 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.877195 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hwmtz"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.878254 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.879555 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-xbjjp"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.880450 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.881110 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dk9k6"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.885164 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-26l22"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.886343 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.888781 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-h2g87"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889189 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889539 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5443f78e-7898-4ee6-9f8f-574331118acd-node-pullsecrets\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889581 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-config\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889613 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889633 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6057de60-fc56-49b5-8843-a9e838186747-config\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889649 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-config\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889665 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-client\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889686 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-oauth-serving-cert\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889726 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh87p\" (UniqueName: \"kubernetes.io/projected/ed93e809-cf53-450c-8993-5a69e97542bd-kube-api-access-xh87p\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889750 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6057de60-fc56-49b5-8843-a9e838186747-images\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889768 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v85pb\" (UniqueName: \"kubernetes.io/projected/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-kube-api-access-v85pb\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889783 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-495kt\" (UniqueName: \"kubernetes.io/projected/a5d04969-439e-47d2-8583-875a3f0ae3d5-kube-api-access-495kt\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889807 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-serving-ca\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889826 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889850 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr2g8\" (UniqueName: \"kubernetes.io/projected/34bc2780-c7b8-45cf-9924-4b08def7fd64-kube-api-access-xr2g8\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889873 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6057de60-fc56-49b5-8843-a9e838186747-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889895 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889914 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-serving-cert\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889935 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889954 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-image-import-ca\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889977 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-client-ca\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889998 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2njm\" (UniqueName: \"kubernetes.io/projected/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-kube-api-access-l2njm\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890018 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-serving-cert\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890039 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-config\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890061 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-encryption-config\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890101 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e908e8d-4252-422b-9d40-f38ef0766dde-serving-cert\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890120 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cb3658-2a8d-49c1-8e8e-a7675d34833c-config\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890154 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-oauth-config\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890177 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jlf2\" (UniqueName: \"kubernetes.io/projected/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-kube-api-access-4jlf2\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890199 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-serving-cert\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890236 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/900dfada-c31e-41ec-97a4-2244cf0d8781-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-49p7f\" (UID: \"900dfada-c31e-41ec-97a4-2244cf0d8781\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890258 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s8tg\" (UniqueName: \"kubernetes.io/projected/6e908e8d-4252-422b-9d40-f38ef0766dde-kube-api-access-4s8tg\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890280 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/20cb3658-2a8d-49c1-8e8e-a7675d34833c-auth-proxy-config\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890295 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890302 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftw5q\" (UniqueName: \"kubernetes.io/projected/900dfada-c31e-41ec-97a4-2244cf0d8781-kube-api-access-ftw5q\") pod \"cluster-samples-operator-665b6dd947-49p7f\" (UID: \"900dfada-c31e-41ec-97a4-2244cf0d8781\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890384 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-config\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890417 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/96e0038e-e6bd-497c-8245-8f11f8428a00-audit-dir\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890462 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890497 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890529 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mxc4\" (UniqueName: \"kubernetes.io/projected/5443f78e-7898-4ee6-9f8f-574331118acd-kube-api-access-2mxc4\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890576 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-etcd-client\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890606 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-audit-policies\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890637 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-config\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890664 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890689 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed93e809-cf53-450c-8993-5a69e97542bd-serving-cert\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890715 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890776 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-ca\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890808 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5d04969-439e-47d2-8583-875a3f0ae3d5-serving-cert\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890857 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gr87\" (UniqueName: \"kubernetes.io/projected/6057de60-fc56-49b5-8843-a9e838186747-kube-api-access-2gr87\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890887 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-encryption-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890917 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-service-ca-bundle\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890950 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-serving-cert\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.890977 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-service-ca\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891003 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-audit\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891030 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34bc2780-c7b8-45cf-9924-4b08def7fd64-serving-cert\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891062 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-trusted-ca\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891092 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6nc6\" (UniqueName: \"kubernetes.io/projected/20cb3658-2a8d-49c1-8e8e-a7675d34833c-kube-api-access-n6nc6\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891118 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v9c4\" (UniqueName: \"kubernetes.io/projected/96e0038e-e6bd-497c-8245-8f11f8428a00-kube-api-access-6v9c4\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891148 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/20cb3658-2a8d-49c1-8e8e-a7675d34833c-machine-approver-tls\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891157 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6057de60-fc56-49b5-8843-a9e838186747-config\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891177 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/34bc2780-c7b8-45cf-9924-4b08def7fd64-available-featuregates\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891250 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-service-ca\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891279 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-trusted-ca-bundle\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891309 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5443f78e-7898-4ee6-9f8f-574331118acd-audit-dir\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891426 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5443f78e-7898-4ee6-9f8f-574331118acd-audit-dir\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.891898 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-config\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.892083 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.892116 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qx6xb"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.892131 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-psj8g"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.892859 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-service-ca-bundle\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.893179 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-service-ca\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.893305 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-trusted-ca\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.893393 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.893864 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-serving-ca\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.894117 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-audit\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.894324 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.894664 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20cb3658-2a8d-49c1-8e8e-a7675d34833c-config\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.897300 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/20cb3658-2a8d-49c1-8e8e-a7675d34833c-machine-approver-tls\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.897419 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a5d04969-439e-47d2-8583-875a3f0ae3d5-serving-cert\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.897625 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2tvsb"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.897654 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.897667 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.897977 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/96e0038e-e6bd-497c-8245-8f11f8428a00-audit-policies\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.898083 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-config\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.898722 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-config\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.898836 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e908e8d-4252-422b-9d40-f38ef0766dde-serving-cert\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.898915 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-etcd-client\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.899263 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34bc2780-c7b8-45cf-9924-4b08def7fd64-serving-cert\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.899721 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-service-ca\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.899898 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-serving-cert\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.899953 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l55b8"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900002 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-config\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900028 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/34bc2780-c7b8-45cf-9924-4b08def7fd64-available-featuregates\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900046 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/96e0038e-e6bd-497c-8245-8f11f8428a00-audit-dir\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900076 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-image-import-ca\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900116 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900666 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-oauth-config\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.900828 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed93e809-cf53-450c-8993-5a69e97542bd-serving-cert\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.901238 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.901305 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-ca\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.901372 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-config\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.889662 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5443f78e-7898-4ee6-9f8f-574331118acd-node-pullsecrets\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.901464 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.901999 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-oauth-serving-cert\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.902767 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-client-ca\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.902835 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.902899 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r7cxf"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903160 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-trusted-ca-bundle\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903338 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/20cb3658-2a8d-49c1-8e8e-a7675d34833c-auth-proxy-config\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903489 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-serving-cert\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903502 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-encryption-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903795 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6057de60-fc56-49b5-8843-a9e838186747-images\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903836 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.903845 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a5d04969-439e-47d2-8583-875a3f0ae3d5-config\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.904155 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6057de60-fc56-49b5-8843-a9e838186747-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.904654 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-encryption-config\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.904786 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.904866 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/900dfada-c31e-41ec-97a4-2244cf0d8781-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-49p7f\" (UID: \"900dfada-c31e-41ec-97a4-2244cf0d8781\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.904920 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/96e0038e-e6bd-497c-8245-8f11f8428a00-serving-cert\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.905323 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.906163 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-serving-cert\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.906277 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-226j2"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.907261 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-ws4lp"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.908281 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.908620 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.909002 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.909874 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-qbhhx"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.910837 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.910838 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.911662 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed93e809-cf53-450c-8993-5a69e97542bd-etcd-client\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.914620 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.914650 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.914660 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-l2x2j"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.916303 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.917259 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.918447 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6w49n"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.919677 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6fsbb"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.931012 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.936151 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.936364 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-psj8g"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.938104 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r77tn"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.939133 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.940167 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.941282 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rnq88"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.942279 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ws4lp"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.943260 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qbhhx"] Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.948628 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.968452 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 07:19:08 crc kubenswrapper[4829]: I1002 07:19:08.989599 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.009151 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.028893 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.048953 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.068945 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093444 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-bound-sa-token\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093495 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmm5k\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-kube-api-access-tmm5k\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093530 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/100e06e7-e40d-4b3f-81cf-86886ad1e928-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093594 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmnmx\" (UniqueName: \"kubernetes.io/projected/100e06e7-e40d-4b3f-81cf-86886ad1e928-kube-api-access-rmnmx\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093651 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/100e06e7-e40d-4b3f-81cf-86886ad1e928-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093769 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093824 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/100e06e7-e40d-4b3f-81cf-86886ad1e928-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093881 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-certificates\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.093918 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-tls\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.094001 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c41ceaa-bc0a-4576-b309-0e7679130b71-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.094042 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.59403092 +0000 UTC m=+140.933679325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.094074 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c41ceaa-bc0a-4576-b309-0e7679130b71-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.094092 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-trusted-ca\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.128865 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.148819 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.168722 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.189258 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.194627 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.194814 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.694766887 +0000 UTC m=+141.034415302 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195183 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9pqw\" (UniqueName: \"kubernetes.io/projected/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-kube-api-access-r9pqw\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195256 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/835adc97-13f4-4100-811b-72d0cc4a2ed3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195333 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwjs2\" (UniqueName: \"kubernetes.io/projected/2cc39185-906e-4b18-a359-b1fd011b74f4-kube-api-access-dwjs2\") pod \"control-plane-machine-set-operator-78cbb6b69f-zqd5d\" (UID: \"2cc39185-906e-4b18-a359-b1fd011b74f4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195374 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/100e06e7-e40d-4b3f-81cf-86886ad1e928-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195412 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195493 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hsxv\" (UniqueName: \"kubernetes.io/projected/ca598dbb-d00b-4361-88c0-34c22ad2be6e-kube-api-access-5hsxv\") pod \"downloads-7954f5f757-r7cxf\" (UID: \"ca598dbb-d00b-4361-88c0-34c22ad2be6e\") " pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195536 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12133a89-e03b-4a3f-b665-c5d790315e98-service-ca-bundle\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195745 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/129ca2ba-597b-40f3-a871-d7196b84b4f0-tmpfs\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.195949 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/129ca2ba-597b-40f3-a871-d7196b84b4f0-apiservice-cert\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.196018 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4xln\" (UniqueName: \"kubernetes.io/projected/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-kube-api-access-q4xln\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.196139 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-socket-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.196486 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.196686 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4481fcdc-864e-4b94-8378-78cb9241dd12-trusted-ca\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.196891 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4481fcdc-864e-4b94-8378-78cb9241dd12-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197024 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-certs\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197124 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-default-certificate\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197300 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d69gk\" (UniqueName: \"kubernetes.io/projected/4481fcdc-864e-4b94-8378-78cb9241dd12-kube-api-access-d69gk\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197412 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxkxd\" (UniqueName: \"kubernetes.io/projected/84c837e4-1328-4abd-ac80-d75637b192d1-kube-api-access-lxkxd\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197533 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt2ws\" (UniqueName: \"kubernetes.io/projected/56c155fb-95c9-46fa-ad70-456dccb31776-kube-api-access-vt2ws\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197579 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197615 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-plugins-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197667 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-tls\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197737 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7dhx\" (UniqueName: \"kubernetes.io/projected/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-kube-api-access-s7dhx\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197773 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-key\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197838 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197873 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-cabundle\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197906 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnmp6\" (UniqueName: \"kubernetes.io/projected/fcb6162b-fb97-4bf5-8563-48c907aa6a36-kube-api-access-fnmp6\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197942 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7b4750a-3c7e-4fa4-9534-40a895624522-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.197976 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-mountpoint-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198012 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c41ceaa-bc0a-4576-b309-0e7679130b71-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198111 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b4750a-3c7e-4fa4-9534-40a895624522-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198149 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198176 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/835adc97-13f4-4100-811b-72d0cc4a2ed3-images\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198274 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dhl5\" (UniqueName: \"kubernetes.io/projected/d5eef982-bef6-4679-be23-88feef0e3793-kube-api-access-4dhl5\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198321 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/129ca2ba-597b-40f3-a871-d7196b84b4f0-webhook-cert\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198343 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t87kd\" (UniqueName: \"kubernetes.io/projected/b1f5af08-068d-43b7-aaa5-45da6c35918c-kube-api-access-t87kd\") pod \"dns-operator-744455d44c-26l22\" (UID: \"b1f5af08-068d-43b7-aaa5-45da6c35918c\") " pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198376 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-policies\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198418 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6rp5\" (UniqueName: \"kubernetes.io/projected/472abadc-66db-4150-9054-9771d488ae44-kube-api-access-n6rp5\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198440 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198472 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-stats-auth\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198494 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-serving-cert\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198515 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6cqh\" (UniqueName: \"kubernetes.io/projected/835adc97-13f4-4100-811b-72d0cc4a2ed3-kube-api-access-v6cqh\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198632 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-bound-sa-token\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198679 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/472abadc-66db-4150-9054-9771d488ae44-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198757 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkjdx\" (UniqueName: \"kubernetes.io/projected/12133a89-e03b-4a3f-b665-c5d790315e98-kube-api-access-nkjdx\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198785 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7b4750a-3c7e-4fa4-9534-40a895624522-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198856 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmm5k\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-kube-api-access-tmm5k\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198879 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2cc39185-906e-4b18-a359-b1fd011b74f4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zqd5d\" (UID: \"2cc39185-906e-4b18-a359-b1fd011b74f4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198954 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b1f5af08-068d-43b7-aaa5-45da6c35918c-metrics-tls\") pod \"dns-operator-744455d44c-26l22\" (UID: \"b1f5af08-068d-43b7-aaa5-45da6c35918c\") " pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.198994 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmnmx\" (UniqueName: \"kubernetes.io/projected/100e06e7-e40d-4b3f-81cf-86886ad1e928-kube-api-access-rmnmx\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199025 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-config-volume\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199067 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-secret-volume\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199095 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4481fcdc-864e-4b94-8378-78cb9241dd12-metrics-tls\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199124 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-csi-data-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199194 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/100e06e7-e40d-4b3f-81cf-86886ad1e928-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199692 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/410e2990-0700-422a-aa64-e425907892c6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199766 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6125f058-c403-48da-87a6-74db78012f40-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dk9k6\" (UID: \"6125f058-c403-48da-87a6-74db78012f40\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199807 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-metrics-certs\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199856 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199895 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-registration-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199931 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.199968 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ps6p\" (UniqueName: \"kubernetes.io/projected/6125f058-c403-48da-87a6-74db78012f40-kube-api-access-9ps6p\") pod \"multus-admission-controller-857f4d67dd-dk9k6\" (UID: \"6125f058-c403-48da-87a6-74db78012f40\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200181 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvt6n\" (UniqueName: \"kubernetes.io/projected/129ca2ba-597b-40f3-a871-d7196b84b4f0-kube-api-access-cvt6n\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200256 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5drbh\" (UniqueName: \"kubernetes.io/projected/410e2990-0700-422a-aa64-e425907892c6-kube-api-access-5drbh\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200313 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/100e06e7-e40d-4b3f-81cf-86886ad1e928-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200349 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/84c837e4-1328-4abd-ac80-d75637b192d1-profile-collector-cert\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.200377 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.700352471 +0000 UTC m=+141.040000946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200552 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c41ceaa-bc0a-4576-b309-0e7679130b71-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200547 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-certificates\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200740 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csz8f\" (UniqueName: \"kubernetes.io/projected/2e0e4181-e7fd-42fc-90b8-5e145f641943-kube-api-access-csz8f\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200773 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/410e2990-0700-422a-aa64-e425907892c6-proxy-tls\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.200819 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/84c837e4-1328-4abd-ac80-d75637b192d1-srv-cert\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201326 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201415 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/835adc97-13f4-4100-811b-72d0cc4a2ed3-proxy-tls\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201564 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-dir\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201736 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5eef982-bef6-4679-be23-88feef0e3793-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201786 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201826 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201893 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlf72\" (UniqueName: \"kubernetes.io/projected/5a36f298-4789-4e89-91a3-d7baef3ed8b2-kube-api-access-wlf72\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201960 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.201997 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-config\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202027 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202112 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c41ceaa-bc0a-4576-b309-0e7679130b71-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202160 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5eef982-bef6-4679-be23-88feef0e3793-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202197 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-trusted-ca\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202252 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202303 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/472abadc-66db-4150-9054-9771d488ae44-srv-cert\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202398 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85zmq\" (UniqueName: \"kubernetes.io/projected/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-kube-api-access-85zmq\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202512 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56c155fb-95c9-46fa-ad70-456dccb31776-cert\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202612 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-certificates\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.202690 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-node-bootstrap-token\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.206646 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/100e06e7-e40d-4b3f-81cf-86886ad1e928-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.206910 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-trusted-ca\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.211142 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.211401 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/100e06e7-e40d-4b3f-81cf-86886ad1e928-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.211681 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c41ceaa-bc0a-4576-b309-0e7679130b71-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.212591 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-tls\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.229862 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.248703 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.268838 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.288414 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.303601 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.303714 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.803691824 +0000 UTC m=+141.143340239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.303927 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjlxd\" (UniqueName: \"kubernetes.io/projected/40922895-36ea-484e-a030-2e9ea196629e-kube-api-access-zjlxd\") pod \"migrator-59844c95c7-4dgbl\" (UID: \"40922895-36ea-484e-a030-2e9ea196629e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.303966 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8706882f-e939-4ac9-b145-527cb52f9294-serving-cert\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.303998 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b4750a-3c7e-4fa4-9534-40a895624522-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304025 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304054 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/835adc97-13f4-4100-811b-72d0cc4a2ed3-images\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304080 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dhl5\" (UniqueName: \"kubernetes.io/projected/d5eef982-bef6-4679-be23-88feef0e3793-kube-api-access-4dhl5\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304103 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0e365b9-9ffa-412b-8055-559480d9d565-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304128 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/129ca2ba-597b-40f3-a871-d7196b84b4f0-webhook-cert\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304151 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t87kd\" (UniqueName: \"kubernetes.io/projected/b1f5af08-068d-43b7-aaa5-45da6c35918c-kube-api-access-t87kd\") pod \"dns-operator-744455d44c-26l22\" (UID: \"b1f5af08-068d-43b7-aaa5-45da6c35918c\") " pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304192 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-policies\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304214 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6rp5\" (UniqueName: \"kubernetes.io/projected/472abadc-66db-4150-9054-9771d488ae44-kube-api-access-n6rp5\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304255 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304278 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-stats-auth\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304299 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6cqh\" (UniqueName: \"kubernetes.io/projected/835adc97-13f4-4100-811b-72d0cc4a2ed3-kube-api-access-v6cqh\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304331 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/472abadc-66db-4150-9054-9771d488ae44-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304353 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-serving-cert\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304375 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkjdx\" (UniqueName: \"kubernetes.io/projected/12133a89-e03b-4a3f-b665-c5d790315e98-kube-api-access-nkjdx\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304400 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7b4750a-3c7e-4fa4-9534-40a895624522-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304422 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-config-volume\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304456 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2cc39185-906e-4b18-a359-b1fd011b74f4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zqd5d\" (UID: \"2cc39185-906e-4b18-a359-b1fd011b74f4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304480 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-metrics-tls\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304514 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b1f5af08-068d-43b7-aaa5-45da6c35918c-metrics-tls\") pod \"dns-operator-744455d44c-26l22\" (UID: \"b1f5af08-068d-43b7-aaa5-45da6c35918c\") " pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304544 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-config-volume\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304566 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-secret-volume\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304588 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4481fcdc-864e-4b94-8378-78cb9241dd12-metrics-tls\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304609 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-csi-data-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304633 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304655 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e365b9-9ffa-412b-8055-559480d9d565-config\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304684 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0e365b9-9ffa-412b-8055-559480d9d565-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304730 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/410e2990-0700-422a-aa64-e425907892c6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304753 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304778 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6125f058-c403-48da-87a6-74db78012f40-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dk9k6\" (UID: \"6125f058-c403-48da-87a6-74db78012f40\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304802 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-metrics-certs\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304830 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304851 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-registration-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304873 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ps6p\" (UniqueName: \"kubernetes.io/projected/6125f058-c403-48da-87a6-74db78012f40-kube-api-access-9ps6p\") pod \"multus-admission-controller-857f4d67dd-dk9k6\" (UID: \"6125f058-c403-48da-87a6-74db78012f40\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304903 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvt6n\" (UniqueName: \"kubernetes.io/projected/129ca2ba-597b-40f3-a871-d7196b84b4f0-kube-api-access-cvt6n\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304925 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304946 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5drbh\" (UniqueName: \"kubernetes.io/projected/410e2990-0700-422a-aa64-e425907892c6-kube-api-access-5drbh\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.304979 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/84c837e4-1328-4abd-ac80-d75637b192d1-profile-collector-cert\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305003 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e639b9b5-75b2-43af-a51c-9402b7dcec03-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jzdnq\" (UID: \"e639b9b5-75b2-43af-a51c-9402b7dcec03\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305034 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d955b93-e95e-4307-9993-07a7f60176bc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305058 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csz8f\" (UniqueName: \"kubernetes.io/projected/2e0e4181-e7fd-42fc-90b8-5e145f641943-kube-api-access-csz8f\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305080 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/410e2990-0700-422a-aa64-e425907892c6-proxy-tls\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305103 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/84c837e4-1328-4abd-ac80-d75637b192d1-srv-cert\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305127 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-config\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305149 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d955b93-e95e-4307-9993-07a7f60176bc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305182 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305205 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/835adc97-13f4-4100-811b-72d0cc4a2ed3-proxy-tls\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305247 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-dir\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305281 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5eef982-bef6-4679-be23-88feef0e3793-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305303 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305325 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-client-ca\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305348 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305370 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305394 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlf72\" (UniqueName: \"kubernetes.io/projected/5a36f298-4789-4e89-91a3-d7baef3ed8b2-kube-api-access-wlf72\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305417 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305441 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-config\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305463 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305487 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5eef982-bef6-4679-be23-88feef0e3793-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305509 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305538 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/472abadc-66db-4150-9054-9771d488ae44-srv-cert\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305560 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85zmq\" (UniqueName: \"kubernetes.io/projected/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-kube-api-access-85zmq\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305582 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56c155fb-95c9-46fa-ad70-456dccb31776-cert\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305606 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-node-bootstrap-token\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305632 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9pqw\" (UniqueName: \"kubernetes.io/projected/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-kube-api-access-r9pqw\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305654 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/835adc97-13f4-4100-811b-72d0cc4a2ed3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305696 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwjs2\" (UniqueName: \"kubernetes.io/projected/2cc39185-906e-4b18-a359-b1fd011b74f4-kube-api-access-dwjs2\") pod \"control-plane-machine-set-operator-78cbb6b69f-zqd5d\" (UID: \"2cc39185-906e-4b18-a359-b1fd011b74f4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305720 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8hfx\" (UniqueName: \"kubernetes.io/projected/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-kube-api-access-w8hfx\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305752 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305776 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hsxv\" (UniqueName: \"kubernetes.io/projected/ca598dbb-d00b-4361-88c0-34c22ad2be6e-kube-api-access-5hsxv\") pod \"downloads-7954f5f757-r7cxf\" (UID: \"ca598dbb-d00b-4361-88c0-34c22ad2be6e\") " pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305808 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/129ca2ba-597b-40f3-a871-d7196b84b4f0-tmpfs\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305831 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12133a89-e03b-4a3f-b665-c5d790315e98-service-ca-bundle\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305854 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/129ca2ba-597b-40f3-a871-d7196b84b4f0-apiservice-cert\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305876 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4xln\" (UniqueName: \"kubernetes.io/projected/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-kube-api-access-q4xln\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305906 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-socket-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305935 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d955b93-e95e-4307-9993-07a7f60176bc-config\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305959 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.305997 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4481fcdc-864e-4b94-8378-78cb9241dd12-trusted-ca\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306021 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc62w\" (UniqueName: \"kubernetes.io/projected/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-kube-api-access-fc62w\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306049 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4481fcdc-864e-4b94-8378-78cb9241dd12-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306070 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-certs\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306090 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-default-certificate\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306114 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d69gk\" (UniqueName: \"kubernetes.io/projected/4481fcdc-864e-4b94-8378-78cb9241dd12-kube-api-access-d69gk\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306138 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxkxd\" (UniqueName: \"kubernetes.io/projected/84c837e4-1328-4abd-ac80-d75637b192d1-kube-api-access-lxkxd\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306164 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt2ws\" (UniqueName: \"kubernetes.io/projected/56c155fb-95c9-46fa-ad70-456dccb31776-kube-api-access-vt2ws\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306189 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j5n7\" (UniqueName: \"kubernetes.io/projected/8706882f-e939-4ac9-b145-527cb52f9294-kube-api-access-7j5n7\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306243 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306267 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-plugins-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306295 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7dhx\" (UniqueName: \"kubernetes.io/projected/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-kube-api-access-s7dhx\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306316 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-key\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306340 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jp4q\" (UniqueName: \"kubernetes.io/projected/e639b9b5-75b2-43af-a51c-9402b7dcec03-kube-api-access-4jp4q\") pod \"package-server-manager-789f6589d5-jzdnq\" (UID: \"e639b9b5-75b2-43af-a51c-9402b7dcec03\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306371 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306394 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-cabundle\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306417 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnmp6\" (UniqueName: \"kubernetes.io/projected/fcb6162b-fb97-4bf5-8563-48c907aa6a36-kube-api-access-fnmp6\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306439 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7b4750a-3c7e-4fa4-9534-40a895624522-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306464 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-mountpoint-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306690 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-mountpoint-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.306932 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-dir\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.307426 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-plugins-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.307473 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-registration-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.307510 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-socket-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.307686 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.807655691 +0000 UTC m=+141.147304096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.308119 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/fcb6162b-fb97-4bf5-8563-48c907aa6a36-csi-data-dir\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.308489 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/12133a89-e03b-4a3f-b665-c5d790315e98-service-ca-bundle\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.308608 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.308647 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/410e2990-0700-422a-aa64-e425907892c6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.309264 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.309717 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4481fcdc-864e-4b94-8378-78cb9241dd12-trusted-ca\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.310042 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/129ca2ba-597b-40f3-a871-d7196b84b4f0-tmpfs\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.311705 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-default-certificate\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.311205 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/835adc97-13f4-4100-811b-72d0cc4a2ed3-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.311871 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.312502 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b1f5af08-068d-43b7-aaa5-45da6c35918c-metrics-tls\") pod \"dns-operator-744455d44c-26l22\" (UID: \"b1f5af08-068d-43b7-aaa5-45da6c35918c\") " pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.315266 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-stats-auth\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.315588 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.315789 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12133a89-e03b-4a3f-b665-c5d790315e98-metrics-certs\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.316069 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4481fcdc-864e-4b94-8378-78cb9241dd12-metrics-tls\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.321067 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.329311 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.339474 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.367399 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.369003 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.373003 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.381356 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.389088 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.402019 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.407513 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.407721 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.907687657 +0000 UTC m=+141.247336102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.407940 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-config-volume\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.407986 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-metrics-tls\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408095 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408130 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e365b9-9ffa-412b-8055-559480d9d565-config\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408172 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408205 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0e365b9-9ffa-412b-8055-559480d9d565-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408323 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408449 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e639b9b5-75b2-43af-a51c-9402b7dcec03-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jzdnq\" (UID: \"e639b9b5-75b2-43af-a51c-9402b7dcec03\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408498 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d955b93-e95e-4307-9993-07a7f60176bc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408566 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-config\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408597 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d955b93-e95e-4307-9993-07a7f60176bc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.408891 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.409153 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-client-ca\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.409208 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:09.9091821 +0000 UTC m=+141.248830545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.409618 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8hfx\" (UniqueName: \"kubernetes.io/projected/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-kube-api-access-w8hfx\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.409685 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0e365b9-9ffa-412b-8055-559480d9d565-config\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.409883 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d955b93-e95e-4307-9993-07a7f60176bc-config\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.410009 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc62w\" (UniqueName: \"kubernetes.io/projected/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-kube-api-access-fc62w\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.409886 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.410276 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j5n7\" (UniqueName: \"kubernetes.io/projected/8706882f-e939-4ac9-b145-527cb52f9294-kube-api-access-7j5n7\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.410399 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jp4q\" (UniqueName: \"kubernetes.io/projected/e639b9b5-75b2-43af-a51c-9402b7dcec03-kube-api-access-4jp4q\") pod \"package-server-manager-789f6589d5-jzdnq\" (UID: \"e639b9b5-75b2-43af-a51c-9402b7dcec03\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.410570 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8706882f-e939-4ac9-b145-527cb52f9294-serving-cert\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.410766 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-config\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.410919 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-client-ca\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.411541 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjlxd\" (UniqueName: \"kubernetes.io/projected/40922895-36ea-484e-a030-2e9ea196629e-kube-api-access-zjlxd\") pod \"migrator-59844c95c7-4dgbl\" (UID: \"40922895-36ea-484e-a030-2e9ea196629e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.411781 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0e365b9-9ffa-412b-8055-559480d9d565-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.412601 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0e365b9-9ffa-412b-8055-559480d9d565-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.412778 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.413462 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d955b93-e95e-4307-9993-07a7f60176bc-config\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.414032 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d955b93-e95e-4307-9993-07a7f60176bc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.415391 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8706882f-e939-4ac9-b145-527cb52f9294-serving-cert\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.474020 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.474100 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.476552 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.483662 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/835adc97-13f4-4100-811b-72d0cc4a2ed3-proxy-tls\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.489070 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.495020 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.512549 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.512802 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.012773121 +0000 UTC m=+141.352421556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.513433 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.513767 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.013750439 +0000 UTC m=+141.353398844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.517853 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.523778 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.529036 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.539077 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.549374 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.555905 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-policies\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.581539 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.589152 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.590727 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.591457 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.610312 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.614941 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.615164 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.115118634 +0000 UTC m=+141.454767079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.615858 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.616899 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.116873666 +0000 UTC m=+141.456522141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.629523 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.649717 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.655965 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/835adc97-13f4-4100-811b-72d0cc4a2ed3-images\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.669774 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.678617 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2cc39185-906e-4b18-a359-b1fd011b74f4-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zqd5d\" (UID: \"2cc39185-906e-4b18-a359-b1fd011b74f4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.689135 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.700284 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/410e2990-0700-422a-aa64-e425907892c6-proxy-tls\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.709087 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.717426 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.717748 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.217711955 +0000 UTC m=+141.557360390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.719065 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.219046774 +0000 UTC m=+141.558695219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.720347 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6125f058-c403-48da-87a6-74db78012f40-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dk9k6\" (UID: \"6125f058-c403-48da-87a6-74db78012f40\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.718628 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.729285 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.749755 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.768303 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.789009 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.809349 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.844456 4829 request.go:700] Waited for 1.032285227s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.844932 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.845270 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.34521963 +0000 UTC m=+141.684868055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.846799 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.851027 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.863847 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5eef982-bef6-4679-be23-88feef0e3793-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.870106 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.890642 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.898928 4829 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.899018 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config podName:5443f78e-7898-4ee6-9f8f-574331118acd nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.398997794 +0000 UTC m=+141.738646209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config") pod "apiserver-76f77b778f-l2x2j" (UID: "5443f78e-7898-4ee6-9f8f-574331118acd") : failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.899422 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/84c837e4-1328-4abd-ac80-d75637b192d1-profile-collector-cert\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.899505 4829 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.899556 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client podName:5443f78e-7898-4ee6-9f8f-574331118acd nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.399539159 +0000 UTC m=+141.739187634 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client") pod "apiserver-76f77b778f-l2x2j" (UID: "5443f78e-7898-4ee6-9f8f-574331118acd") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.909296 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-secret-volume\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.909775 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/472abadc-66db-4150-9054-9771d488ae44-profile-collector-cert\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.910732 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.918684 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/84c837e4-1328-4abd-ac80-d75637b192d1-srv-cert\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.929536 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.947479 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:09 crc kubenswrapper[4829]: E1002 07:19:09.947912 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.447889163 +0000 UTC m=+141.787537638 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.949398 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.960365 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5eef982-bef6-4679-be23-88feef0e3793-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.969544 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 07:19:09 crc kubenswrapper[4829]: I1002 07:19:09.989782 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.008755 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.029121 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.043363 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e639b9b5-75b2-43af-a51c-9402b7dcec03-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jzdnq\" (UID: \"e639b9b5-75b2-43af-a51c-9402b7dcec03\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.048653 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.048725 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.048858 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.548823725 +0000 UTC m=+141.888472130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.049155 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.049660 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.549639039 +0000 UTC m=+141.889287484 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.055110 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/472abadc-66db-4150-9054-9771d488ae44-srv-cert\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.072086 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.079997 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/129ca2ba-597b-40f3-a871-d7196b84b4f0-webhook-cert\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.081218 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/129ca2ba-597b-40f3-a871-d7196b84b4f0-apiservice-cert\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.090414 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.109369 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.120390 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-config-volume\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.130324 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.150749 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.151000 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.650971554 +0000 UTC m=+141.990619959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.151429 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.151788 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.152308 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.652274071 +0000 UTC m=+141.991922546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.179263 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.189852 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.190735 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.202423 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.209630 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.230343 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.238364 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7b4750a-3c7e-4fa4-9534-40a895624522-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.249283 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.253415 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.253582 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.753551363 +0000 UTC m=+142.093199808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.254289 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.254797 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.75478084 +0000 UTC m=+142.094429285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.258587 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7b4750a-3c7e-4fa4-9534-40a895624522-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.269878 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.288789 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.305401 4829 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.305512 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-serving-cert podName:3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.805485332 +0000 UTC m=+142.145133777 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-serving-cert") pod "service-ca-operator-777779d784-mcxkp" (UID: "3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.307289 4829 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.307353 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56c155fb-95c9-46fa-ad70-456dccb31776-cert podName:56c155fb-95c9-46fa-ad70-456dccb31776 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.807334887 +0000 UTC m=+142.146983302 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/56c155fb-95c9-46fa-ad70-456dccb31776-cert") pod "ingress-canary-ws4lp" (UID: "56c155fb-95c9-46fa-ad70-456dccb31776") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.307791 4829 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.307905 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-cabundle podName:5a36f298-4789-4e89-91a3-d7baef3ed8b2 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.807880483 +0000 UTC m=+142.147528918 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-cabundle") pod "service-ca-9c57cc56f-6fsbb" (UID: "5a36f298-4789-4e89-91a3-d7baef3ed8b2") : failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308563 4829 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308610 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-config podName:3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.808599095 +0000 UTC m=+142.148247509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-config") pod "service-ca-operator-777779d784-mcxkp" (UID: "3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517") : failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308781 4829 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308892 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-node-bootstrap-token podName:dddb80c0-70db-4c01-8bfb-ce39a6cd1406 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.808865243 +0000 UTC m=+142.148513718 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-node-bootstrap-token") pod "machine-config-server-xbjjp" (UID: "dddb80c0-70db-4c01-8bfb-ce39a6cd1406") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308796 4829 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.308935 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308946 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-certs podName:dddb80c0-70db-4c01-8bfb-ce39a6cd1406 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.808937125 +0000 UTC m=+142.148585540 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-certs") pod "machine-config-server-xbjjp" (UID: "dddb80c0-70db-4c01-8bfb-ce39a6cd1406") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.308937 4829 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.309077 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-key podName:5a36f298-4789-4e89-91a3-d7baef3ed8b2 nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.809053908 +0000 UTC m=+142.148702483 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-key") pod "service-ca-9c57cc56f-6fsbb" (UID: "5a36f298-4789-4e89-91a3-d7baef3ed8b2") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.329277 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.349344 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.355879 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.356022 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.85599314 +0000 UTC m=+142.195641545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.356898 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.357521 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.857510975 +0000 UTC m=+142.197159380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.369527 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.388872 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.408527 4829 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.408667 4829 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.408764 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-config-volume podName:4b0294ed-4a71-4a99-abf6-170b7ba35d7c nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.908738793 +0000 UTC m=+142.248387238 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-config-volume") pod "dns-default-qbhhx" (UID: "4b0294ed-4a71-4a99-abf6-170b7ba35d7c") : failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.408778 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.409051 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-metrics-tls podName:4b0294ed-4a71-4a99-abf6-170b7ba35d7c nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.909009271 +0000 UTC m=+142.248657726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-metrics-tls") pod "dns-default-qbhhx" (UID: "4b0294ed-4a71-4a99-abf6-170b7ba35d7c") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.428535 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.448982 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.463494 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.463922 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.963753214 +0000 UTC m=+142.303401609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.465690 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.465772 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.465818 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.466659 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:10.966624487 +0000 UTC m=+142.306272932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.470878 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.489622 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.509864 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.528577 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.549653 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.567329 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.567853 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.067823168 +0000 UTC m=+142.407471613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.568464 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.568948 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.06892715 +0000 UTC m=+142.408575585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.598134 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftw5q\" (UniqueName: \"kubernetes.io/projected/900dfada-c31e-41ec-97a4-2244cf0d8781-kube-api-access-ftw5q\") pod \"cluster-samples-operator-665b6dd947-49p7f\" (UID: \"900dfada-c31e-41ec-97a4-2244cf0d8781\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.616661 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gr87\" (UniqueName: \"kubernetes.io/projected/6057de60-fc56-49b5-8843-a9e838186747-kube-api-access-2gr87\") pod \"machine-api-operator-5694c8668f-hwmtz\" (UID: \"6057de60-fc56-49b5-8843-a9e838186747\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.629325 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.635805 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr2g8\" (UniqueName: \"kubernetes.io/projected/34bc2780-c7b8-45cf-9924-4b08def7fd64-kube-api-access-xr2g8\") pod \"openshift-config-operator-7777fb866f-226j2\" (UID: \"34bc2780-c7b8-45cf-9924-4b08def7fd64\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.649427 4829 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.669343 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.669492 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.669648 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.169612704 +0000 UTC m=+142.509261149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.670006 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.670516 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.170495771 +0000 UTC m=+142.510144206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.714446 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-495kt\" (UniqueName: \"kubernetes.io/projected/a5d04969-439e-47d2-8583-875a3f0ae3d5-kube-api-access-495kt\") pod \"authentication-operator-69f744f599-l55b8\" (UID: \"a5d04969-439e-47d2-8583-875a3f0ae3d5\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.729097 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6nc6\" (UniqueName: \"kubernetes.io/projected/20cb3658-2a8d-49c1-8e8e-a7675d34833c-kube-api-access-n6nc6\") pod \"machine-approver-56656f9798-br5pp\" (UID: \"20cb3658-2a8d-49c1-8e8e-a7675d34833c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.757294 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.761014 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v9c4\" (UniqueName: \"kubernetes.io/projected/96e0038e-e6bd-497c-8245-8f11f8428a00-kube-api-access-6v9c4\") pod \"apiserver-7bbb656c7d-wfrrj\" (UID: \"96e0038e-e6bd-497c-8245-8f11f8428a00\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.763301 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.771735 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.771894 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.271861626 +0000 UTC m=+142.611510071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.772038 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.772788 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.272771353 +0000 UTC m=+142.612419798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.792179 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jlf2\" (UniqueName: \"kubernetes.io/projected/7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979-kube-api-access-4jlf2\") pod \"console-operator-58897d9998-rnq88\" (UID: \"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979\") " pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.797345 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh87p\" (UniqueName: \"kubernetes.io/projected/ed93e809-cf53-450c-8993-5a69e97542bd-kube-api-access-xh87p\") pod \"etcd-operator-b45778765-h2g87\" (UID: \"ed93e809-cf53-450c-8993-5a69e97542bd\") " pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.816425 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mxc4\" (UniqueName: \"kubernetes.io/projected/5443f78e-7898-4ee6-9f8f-574331118acd-kube-api-access-2mxc4\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.836988 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s8tg\" (UniqueName: \"kubernetes.io/projected/6e908e8d-4252-422b-9d40-f38ef0766dde-kube-api-access-4s8tg\") pod \"route-controller-manager-6576b87f9c-7d9w8\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.847277 4829 request.go:700] Waited for 1.944333592s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.853935 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2njm\" (UniqueName: \"kubernetes.io/projected/9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a-kube-api-access-l2njm\") pod \"openshift-apiserver-operator-796bbdcf4f-tjn2t\" (UID: \"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.856648 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.869024 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v85pb\" (UniqueName: \"kubernetes.io/projected/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-kube-api-access-v85pb\") pod \"console-f9d7485db-2tvsb\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.869370 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874287 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.874453 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.374429795 +0000 UTC m=+142.714078220 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874661 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-key\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874689 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-cabundle\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874760 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-serving-cert\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874838 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874910 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-config\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874933 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56c155fb-95c9-46fa-ad70-456dccb31776-cert\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.874953 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-node-bootstrap-token\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.875023 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-certs\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.875436 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.375428176 +0000 UTC m=+142.715076581 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.876341 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-config\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.878128 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-cabundle\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.879084 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-certs\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.879803 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/5a36f298-4789-4e89-91a3-d7baef3ed8b2-signing-key\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.880550 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-serving-cert\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.887942 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-node-bootstrap-token\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.889909 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.901415 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56c155fb-95c9-46fa-ad70-456dccb31776-cert\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.910527 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.929871 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.950307 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.950888 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.964915 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.969371 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.970592 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.975384 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.976160 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.976352 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.476334347 +0000 UTC m=+142.815982752 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.976512 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-config-volume\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.976544 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-metrics-tls\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.976609 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:10 crc kubenswrapper[4829]: E1002 07:19:10.977042 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.477024007 +0000 UTC m=+142.816672412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.979628 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-config-volume\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:10 crc kubenswrapper[4829]: I1002 07:19:10.989516 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.007738 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-metrics-tls\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.035506 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.044428 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.048359 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/100e06e7-e40d-4b3f-81cf-86886ad1e928-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.049655 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.068984 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.074969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-bound-sa-token\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.077423 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.077728 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.577703061 +0000 UTC m=+142.917351466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.084675 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmnmx\" (UniqueName: \"kubernetes.io/projected/100e06e7-e40d-4b3f-81cf-86886ad1e928-kube-api-access-rmnmx\") pod \"cluster-image-registry-operator-dc59b4c8b-kcs46\" (UID: \"100e06e7-e40d-4b3f-81cf-86886ad1e928\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.107514 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmm5k\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-kube-api-access-tmm5k\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.114447 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-hwmtz"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.126308 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dhl5\" (UniqueName: \"kubernetes.io/projected/d5eef982-bef6-4679-be23-88feef0e3793-kube-api-access-4dhl5\") pod \"kube-storage-version-migrator-operator-b67b599dd-9v6qs\" (UID: \"d5eef982-bef6-4679-be23-88feef0e3793\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.155525 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t87kd\" (UniqueName: \"kubernetes.io/projected/b1f5af08-068d-43b7-aaa5-45da6c35918c-kube-api-access-t87kd\") pod \"dns-operator-744455d44c-26l22\" (UID: \"b1f5af08-068d-43b7-aaa5-45da6c35918c\") " pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.163942 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-26l22" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.176135 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6rp5\" (UniqueName: \"kubernetes.io/projected/472abadc-66db-4150-9054-9771d488ae44-kube-api-access-n6rp5\") pod \"olm-operator-6b444d44fb-rjdl6\" (UID: \"472abadc-66db-4150-9054-9771d488ae44\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.183909 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.184583 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.684568968 +0000 UTC m=+143.024217373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.186782 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.189666 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.191282 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6cqh\" (UniqueName: \"kubernetes.io/projected/835adc97-13f4-4100-811b-72d0cc4a2ed3-kube-api-access-v6cqh\") pod \"machine-config-operator-74547568cd-6jt2q\" (UID: \"835adc97-13f4-4100-811b-72d0cc4a2ed3\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.195995 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.201620 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkjdx\" (UniqueName: \"kubernetes.io/projected/12133a89-e03b-4a3f-b665-c5d790315e98-kube-api-access-nkjdx\") pod \"router-default-5444994796-cmbx8\" (UID: \"12133a89-e03b-4a3f-b665-c5d790315e98\") " pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:11 crc kubenswrapper[4829]: W1002 07:19:11.224716 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96e0038e_e6bd_497c_8245_8f11f8428a00.slice/crio-00d35cf3ce3c3684ecead1745ee8beba57dac271e362a650acbb9c77eefc02cf WatchSource:0}: Error finding container 00d35cf3ce3c3684ecead1745ee8beba57dac271e362a650acbb9c77eefc02cf: Status 404 returned error can't find the container with id 00d35cf3ce3c3684ecead1745ee8beba57dac271e362a650acbb9c77eefc02cf Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.226637 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7b4750a-3c7e-4fa4-9534-40a895624522-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-vm74c\" (UID: \"e7b4750a-3c7e-4fa4-9534-40a895624522\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.231795 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.236910 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-226j2"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.244648 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5drbh\" (UniqueName: \"kubernetes.io/projected/410e2990-0700-422a-aa64-e425907892c6-kube-api-access-5drbh\") pod \"machine-config-controller-84d6567774-8nnbv\" (UID: \"410e2990-0700-422a-aa64-e425907892c6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.252164 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-l55b8"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.253088 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.264161 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csz8f\" (UniqueName: \"kubernetes.io/projected/2e0e4181-e7fd-42fc-90b8-5e145f641943-kube-api-access-csz8f\") pod \"oauth-openshift-558db77b4-6w49n\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.285755 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.285989 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:11 crc kubenswrapper[4829]: W1002 07:19:11.286745 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34bc2780_c7b8_45cf_9924_4b08def7fd64.slice/crio-806404efd3292baad5f9b7dba48503bd8bfa5337c667eec14893381b7fc7dd66 WatchSource:0}: Error finding container 806404efd3292baad5f9b7dba48503bd8bfa5337c667eec14893381b7fc7dd66: Status 404 returned error can't find the container with id 806404efd3292baad5f9b7dba48503bd8bfa5337c667eec14893381b7fc7dd66 Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.293835 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d69gk\" (UniqueName: \"kubernetes.io/projected/4481fcdc-864e-4b94-8378-78cb9241dd12-kube-api-access-d69gk\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.294396 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.792169516 +0000 UTC m=+143.131817921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.301957 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.310660 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxkxd\" (UniqueName: \"kubernetes.io/projected/84c837e4-1328-4abd-ac80-d75637b192d1-kube-api-access-lxkxd\") pod \"catalog-operator-68c6474976-4dk4r\" (UID: \"84c837e4-1328-4abd-ac80-d75637b192d1\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.341198 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt2ws\" (UniqueName: \"kubernetes.io/projected/56c155fb-95c9-46fa-ad70-456dccb31776-kube-api-access-vt2ws\") pod \"ingress-canary-ws4lp\" (UID: \"56c155fb-95c9-46fa-ad70-456dccb31776\") " pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.341749 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2tvsb"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.351775 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlf72\" (UniqueName: \"kubernetes.io/projected/5a36f298-4789-4e89-91a3-d7baef3ed8b2-kube-api-access-wlf72\") pod \"service-ca-9c57cc56f-6fsbb\" (UID: \"5a36f298-4789-4e89-91a3-d7baef3ed8b2\") " pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.371288 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-h2g87"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.371520 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnmp6\" (UniqueName: \"kubernetes.io/projected/fcb6162b-fb97-4bf5-8563-48c907aa6a36-kube-api-access-fnmp6\") pod \"csi-hostpathplugin-psj8g\" (UID: \"fcb6162b-fb97-4bf5-8563-48c907aa6a36\") " pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.382472 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.386433 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hsxv\" (UniqueName: \"kubernetes.io/projected/ca598dbb-d00b-4361-88c0-34c22ad2be6e-kube-api-access-5hsxv\") pod \"downloads-7954f5f757-r7cxf\" (UID: \"ca598dbb-d00b-4361-88c0-34c22ad2be6e\") " pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.387505 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.387939 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.887924197 +0000 UTC m=+143.227572602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.389187 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.404199 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.408645 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.412726 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7dhx\" (UniqueName: \"kubernetes.io/projected/3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517-kube-api-access-s7dhx\") pod \"service-ca-operator-777779d784-mcxkp\" (UID: \"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.413342 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.428337 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4xln\" (UniqueName: \"kubernetes.io/projected/72cf0023-3aa0-4ee6-9696-fb3155c8f8c2-kube-api-access-q4xln\") pod \"openshift-controller-manager-operator-756b6f6bc6-mpm4h\" (UID: \"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.430730 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.431625 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" event={"ID":"20cb3658-2a8d-49c1-8e8e-a7675d34833c","Type":"ContainerStarted","Data":"0d976c16cb16d22bf3e23923510e51ef6f39ed606bedc4f128af563b9da239c7"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.436252 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" event={"ID":"34bc2780-c7b8-45cf-9924-4b08def7fd64","Type":"ContainerStarted","Data":"806404efd3292baad5f9b7dba48503bd8bfa5337c667eec14893381b7fc7dd66"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.437318 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" event={"ID":"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a","Type":"ContainerStarted","Data":"6522e49d5245906b55b04509a7d593f8a07a6046a97abb0ddfcb4d8edb3548e4"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.438501 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" event={"ID":"6057de60-fc56-49b5-8843-a9e838186747","Type":"ContainerStarted","Data":"f042ed5f7ad554e66de6b2e33c77012cee39b95c66219b996310ff0edbb32178"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.438520 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" event={"ID":"6057de60-fc56-49b5-8843-a9e838186747","Type":"ContainerStarted","Data":"edf4eedad3c1dbfc399546665ce26895c810fc6c0a83ad069b40fc4a556a1479"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.439115 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" event={"ID":"a5d04969-439e-47d2-8583-875a3f0ae3d5","Type":"ContainerStarted","Data":"a80c2e9a6b7808d266206a4050bd9472eb518dece1b49d594ba240ce3805d45e"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.439670 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" event={"ID":"96e0038e-e6bd-497c-8245-8f11f8428a00","Type":"ContainerStarted","Data":"00d35cf3ce3c3684ecead1745ee8beba57dac271e362a650acbb9c77eefc02cf"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.440192 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" event={"ID":"6e908e8d-4252-422b-9d40-f38ef0766dde","Type":"ContainerStarted","Data":"b05012b089903b3c8e30de0b3526d9a8815550c2fd7a004db467c17249e97f24"} Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.462298 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwjs2\" (UniqueName: \"kubernetes.io/projected/2cc39185-906e-4b18-a359-b1fd011b74f4-kube-api-access-dwjs2\") pod \"control-plane-machine-set-operator-78cbb6b69f-zqd5d\" (UID: \"2cc39185-906e-4b18-a359-b1fd011b74f4\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.464135 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.467426 4829 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.467482 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config podName:5443f78e-7898-4ee6-9f8f-574331118acd nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.467464799 +0000 UTC m=+143.807113204 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config") pod "apiserver-76f77b778f-l2x2j" (UID: "5443f78e-7898-4ee6-9f8f-574331118acd") : failed to sync configmap cache: timed out waiting for the condition Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.467788 4829 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.467815 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client podName:5443f78e-7898-4ee6-9f8f-574331118acd nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.467809189 +0000 UTC m=+143.807457594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client") pod "apiserver-76f77b778f-l2x2j" (UID: "5443f78e-7898-4ee6-9f8f-574331118acd") : failed to sync secret cache: timed out waiting for the condition Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.472666 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.479557 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ps6p\" (UniqueName: \"kubernetes.io/projected/6125f058-c403-48da-87a6-74db78012f40-kube-api-access-9ps6p\") pod \"multus-admission-controller-857f4d67dd-dk9k6\" (UID: \"6125f058-c403-48da-87a6-74db78012f40\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.488102 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.488647 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.988266841 +0000 UTC m=+143.327915236 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.488990 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.489440 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:11.989431935 +0000 UTC m=+143.329080340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.491283 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9pqw\" (UniqueName: \"kubernetes.io/projected/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-kube-api-access-r9pqw\") pod \"collect-profiles-29323155-g97qs\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.491909 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.501403 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4481fcdc-864e-4b94-8378-78cb9241dd12-bound-sa-token\") pod \"ingress-operator-5b745b69d9-f68cm\" (UID: \"4481fcdc-864e-4b94-8378-78cb9241dd12\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.516050 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.526291 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvt6n\" (UniqueName: \"kubernetes.io/projected/129ca2ba-597b-40f3-a871-d7196b84b4f0-kube-api-access-cvt6n\") pod \"packageserver-d55dfcdfc-ghvdx\" (UID: \"129ca2ba-597b-40f3-a871-d7196b84b4f0\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.546606 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.548063 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-26l22"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.550924 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.551607 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85zmq\" (UniqueName: \"kubernetes.io/projected/dddb80c0-70db-4c01-8bfb-ce39a6cd1406-kube-api-access-85zmq\") pod \"machine-config-server-xbjjp\" (UID: \"dddb80c0-70db-4c01-8bfb-ce39a6cd1406\") " pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.564814 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-xbjjp" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.565332 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d955b93-e95e-4307-9993-07a7f60176bc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kzcst\" (UID: \"4d955b93-e95e-4307-9993-07a7f60176bc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.580403 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rnq88"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.627433 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-ws4lp" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.627474 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.627556 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.627954 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.127938214 +0000 UTC m=+143.467586619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.630670 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8hfx\" (UniqueName: \"kubernetes.io/projected/4b0294ed-4a71-4a99-abf6-170b7ba35d7c-kube-api-access-w8hfx\") pod \"dns-default-qbhhx\" (UID: \"4b0294ed-4a71-4a99-abf6-170b7ba35d7c\") " pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.631072 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.651389 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j5n7\" (UniqueName: \"kubernetes.io/projected/8706882f-e939-4ac9-b145-527cb52f9294-kube-api-access-7j5n7\") pod \"controller-manager-879f6c89f-tsrrk\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.652052 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc62w\" (UniqueName: \"kubernetes.io/projected/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-kube-api-access-fc62w\") pod \"marketplace-operator-79b997595-r77tn\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.653138 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjlxd\" (UniqueName: \"kubernetes.io/projected/40922895-36ea-484e-a030-2e9ea196629e-kube-api-access-zjlxd\") pod \"migrator-59844c95c7-4dgbl\" (UID: \"40922895-36ea-484e-a030-2e9ea196629e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.672653 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0e365b9-9ffa-412b-8055-559480d9d565-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ctfz2\" (UID: \"d0e365b9-9ffa-412b-8055-559480d9d565\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.676295 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.684448 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jp4q\" (UniqueName: \"kubernetes.io/projected/e639b9b5-75b2-43af-a51c-9402b7dcec03-kube-api-access-4jp4q\") pod \"package-server-manager-789f6589d5-jzdnq\" (UID: \"e639b9b5-75b2-43af-a51c-9402b7dcec03\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.690006 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.695874 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.709936 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 07:19:11 crc kubenswrapper[4829]: W1002 07:19:11.713903 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a9ac0f1_deb2_4be2_9aa2_a2bf69c7d979.slice/crio-f511f09038ae117c874dd44aa636f3970d428b5888b7832ff628ea198d287698 WatchSource:0}: Error finding container f511f09038ae117c874dd44aa636f3970d428b5888b7832ff628ea198d287698: Status 404 returned error can't find the container with id f511f09038ae117c874dd44aa636f3970d428b5888b7832ff628ea198d287698 Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.722366 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.741248 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.741595 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.24158354 +0000 UTC m=+143.581231945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.742083 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.742491 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.742778 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.749639 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.783543 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.794436 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.802332 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.809687 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.811701 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs"] Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.826638 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.841906 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.842298 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.342284505 +0000 UTC m=+143.681932910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:11 crc kubenswrapper[4829]: W1002 07:19:11.876434 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd5eef982_bef6_4679_be23_88feef0e3793.slice/crio-6cbac941747f6c02149508d521cc223866ea82263029ea3d56902ff1d09839b9 WatchSource:0}: Error finding container 6cbac941747f6c02149508d521cc223866ea82263029ea3d56902ff1d09839b9: Status 404 returned error can't find the container with id 6cbac941747f6c02149508d521cc223866ea82263029ea3d56902ff1d09839b9 Oct 02 07:19:11 crc kubenswrapper[4829]: I1002 07:19:11.942945 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:11 crc kubenswrapper[4829]: E1002 07:19:11.943289 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.443272959 +0000 UTC m=+143.782921364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.050075 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.050639 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.55062208 +0000 UTC m=+143.890270485 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.079200 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-r7cxf"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.104812 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.124576 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6w49n"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.151797 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.152146 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.652134859 +0000 UTC m=+143.991783264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.161115 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.252679 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.253111 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.753096222 +0000 UTC m=+144.092744627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: W1002 07:19:12.291368 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod410e2990_0700_422a_aa64_e425907892c6.slice/crio-5d7d35a573c6b85514580ca01f52847d5f8e9a7d8567785c8d0fd072e95e6492 WatchSource:0}: Error finding container 5d7d35a573c6b85514580ca01f52847d5f8e9a7d8567785c8d0fd072e95e6492: Status 404 returned error can't find the container with id 5d7d35a573c6b85514580ca01f52847d5f8e9a7d8567785c8d0fd072e95e6492 Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.355264 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.355781 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.855765845 +0000 UTC m=+144.195414250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.461368 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.462868 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:12.962842088 +0000 UTC m=+144.302490503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.465399 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r7cxf" event={"ID":"ca598dbb-d00b-4361-88c0-34c22ad2be6e","Type":"ContainerStarted","Data":"19146d7351ecbc29656cf4252db11c2e9befbeb3502ab8113e63dea0dc2b6f9b"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.467828 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" event={"ID":"e7b4750a-3c7e-4fa4-9534-40a895624522","Type":"ContainerStarted","Data":"42cd6e7219a993667999a9d0b5e9eb2a95d4d6b60c3ccf0b35ec5350060d583e"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.473673 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" event={"ID":"ed93e809-cf53-450c-8993-5a69e97542bd","Type":"ContainerStarted","Data":"d23a3df8f08311fe0a8528370e4110d2314b15b5765d7608ec33c25a456505ee"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.476598 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" event={"ID":"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517","Type":"ContainerStarted","Data":"b384523fe8c2d945d4aedce410263d094cc9026346e45e31a72f89eee3a388cb"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.480715 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2tvsb" event={"ID":"4c7291f2-090f-4ab6-bcd5-203fee2bbce1","Type":"ContainerStarted","Data":"5cfcc4ee06baacdcb73ca3effabc7e025cfab177e1da0eb52631f12fb70752b6"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.484153 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" event={"ID":"d5eef982-bef6-4679-be23-88feef0e3793","Type":"ContainerStarted","Data":"6cbac941747f6c02149508d521cc223866ea82263029ea3d56902ff1d09839b9"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.486812 4829 generic.go:334] "Generic (PLEG): container finished" podID="96e0038e-e6bd-497c-8245-8f11f8428a00" containerID="76f89116d4294f8333df316ebeb73f6d04b9e4bc35d597acc0bfad6ec3c3c52b" exitCode=0 Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.486870 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" event={"ID":"96e0038e-e6bd-497c-8245-8f11f8428a00","Type":"ContainerDied","Data":"76f89116d4294f8333df316ebeb73f6d04b9e4bc35d597acc0bfad6ec3c3c52b"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.487685 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rnq88" event={"ID":"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979","Type":"ContainerStarted","Data":"f511f09038ae117c874dd44aa636f3970d428b5888b7832ff628ea198d287698"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.488682 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xbjjp" event={"ID":"dddb80c0-70db-4c01-8bfb-ce39a6cd1406","Type":"ContainerStarted","Data":"50d02574a735df007d0c2e30954f4fec83a9d93cce3a5f93390f06abe1095b41"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.490543 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" event={"ID":"20cb3658-2a8d-49c1-8e8e-a7675d34833c","Type":"ContainerStarted","Data":"658a90413cf3e19f445a85ccf571d8f5929da85627678f3a60054271998198aa"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.492463 4829 generic.go:334] "Generic (PLEG): container finished" podID="34bc2780-c7b8-45cf-9924-4b08def7fd64" containerID="e5ecbb515e9717d793103ea6c82b0b8421ea3c86244bb28b4b57e7792228f4ae" exitCode=0 Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.492626 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" event={"ID":"34bc2780-c7b8-45cf-9924-4b08def7fd64","Type":"ContainerDied","Data":"e5ecbb515e9717d793103ea6c82b0b8421ea3c86244bb28b4b57e7792228f4ae"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.493822 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cmbx8" event={"ID":"12133a89-e03b-4a3f-b665-c5d790315e98","Type":"ContainerStarted","Data":"f88b4195d4bfcdd860a510655abce75a23a35ddd354ed3b4017ed99438417d77"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.495417 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" event={"ID":"6057de60-fc56-49b5-8843-a9e838186747","Type":"ContainerStarted","Data":"e3a20e615048650ba83075b5f36ac5d51767cc9850b9d85818fd58f0ebf5d697"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.497099 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" event={"ID":"a5d04969-439e-47d2-8583-875a3f0ae3d5","Type":"ContainerStarted","Data":"545c11cfbfe9a7e2ceac7a4d9e599225cf2d7349463f061caecee1989008757a"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.501304 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" event={"ID":"410e2990-0700-422a-aa64-e425907892c6","Type":"ContainerStarted","Data":"5d7d35a573c6b85514580ca01f52847d5f8e9a7d8567785c8d0fd072e95e6492"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.507547 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-26l22" event={"ID":"b1f5af08-068d-43b7-aaa5-45da6c35918c","Type":"ContainerStarted","Data":"8e29e07cbb516519bba2184e92d9451a3c8eff185ee1797c7e32cdd34c3ff998"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.519263 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" event={"ID":"9aca3a6c-ce2d-4ff8-b82a-44aee2bbe81a","Type":"ContainerStarted","Data":"46a2b7902d5cf69d556f425a2d16f03d4d859f860cf1bf4c3ffd0468a69fc199"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.520014 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" event={"ID":"100e06e7-e40d-4b3f-81cf-86886ad1e928","Type":"ContainerStarted","Data":"e1d91458ed91838a320faff104cac2d4300d4d796bd37879edf147d084475f48"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.521000 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" event={"ID":"472abadc-66db-4150-9054-9771d488ae44","Type":"ContainerStarted","Data":"a6b2e7aab1e3e7cfb77326597cafa53db9670c7a6ab40ad6bb4340b3b9ba5bb5"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.522788 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" event={"ID":"6e908e8d-4252-422b-9d40-f38ef0766dde","Type":"ContainerStarted","Data":"f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.523370 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.524419 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" event={"ID":"2e0e4181-e7fd-42fc-90b8-5e145f641943","Type":"ContainerStarted","Data":"d9b490ca241d1556407e0500828829569d03aa8755918c9eb18666021b636743"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.531036 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" event={"ID":"900dfada-c31e-41ec-97a4-2244cf0d8781","Type":"ContainerStarted","Data":"d23a0db125f098b4584bb5b6a80e9f15a81f0e675ee5ff86e0d110c964697707"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.531101 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" event={"ID":"900dfada-c31e-41ec-97a4-2244cf0d8781","Type":"ContainerStarted","Data":"cd905e4ecac64d0fa2f38e9e0cf5dba91f9567c09a397e1587df2ddfc4f2aa9b"} Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.565917 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.565979 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.566005 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.569317 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.069303053 +0000 UTC m=+144.408951458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.569641 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5443f78e-7898-4ee6-9f8f-574331118acd-config\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.590395 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5443f78e-7898-4ee6-9f8f-574331118acd-etcd-client\") pod \"apiserver-76f77b778f-l2x2j\" (UID: \"5443f78e-7898-4ee6-9f8f-574331118acd\") " pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.603527 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.616770 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6fsbb"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.620823 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.644099 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.657095 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.669709 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.671236 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.171181062 +0000 UTC m=+144.510829467 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.676393 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-psj8g"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.736271 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.772472 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.773003 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.27296476 +0000 UTC m=+144.612613165 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.872714 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-qbhhx"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.873215 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.873808 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.373769318 +0000 UTC m=+144.713417723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.888476 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h"] Oct 02 07:19:12 crc kubenswrapper[4829]: I1002 07:19:12.974648 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:12 crc kubenswrapper[4829]: E1002 07:19:12.975520 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.475505724 +0000 UTC m=+144.815154119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: W1002 07:19:13.011441 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72cf0023_3aa0_4ee6_9696_fb3155c8f8c2.slice/crio-06cbe01a688235cc264de23699b57a13557bfc09bc6ef21f6c9f38fa8173b9e5 WatchSource:0}: Error finding container 06cbe01a688235cc264de23699b57a13557bfc09bc6ef21f6c9f38fa8173b9e5: Status 404 returned error can't find the container with id 06cbe01a688235cc264de23699b57a13557bfc09bc6ef21f6c9f38fa8173b9e5 Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.061025 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-ws4lp"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.077825 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.077987 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.577965071 +0000 UTC m=+144.917613476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.078119 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.080369 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.580348291 +0000 UTC m=+144.919996896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.096483 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dk9k6"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.111994 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d"] Oct 02 07:19:13 crc kubenswrapper[4829]: W1002 07:19:13.114196 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d955b93_e95e_4307_9993_07a7f60176bc.slice/crio-9eb532d0ada507714791b473239610fc3e9b801bff6d09acf7d857665b0291f8 WatchSource:0}: Error finding container 9eb532d0ada507714791b473239610fc3e9b801bff6d09acf7d857665b0291f8: Status 404 returned error can't find the container with id 9eb532d0ada507714791b473239610fc3e9b801bff6d09acf7d857665b0291f8 Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.120373 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.131879 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.138730 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsrrk"] Oct 02 07:19:13 crc kubenswrapper[4829]: W1002 07:19:13.144313 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4481fcdc_864e_4b94_8378_78cb9241dd12.slice/crio-13da5ad920745f5a116b927b366cd3a00517e171af0951106a2e8183153b7d0d WatchSource:0}: Error finding container 13da5ad920745f5a116b927b366cd3a00517e171af0951106a2e8183153b7d0d: Status 404 returned error can't find the container with id 13da5ad920745f5a116b927b366cd3a00517e171af0951106a2e8183153b7d0d Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.151682 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.159854 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.161852 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r77tn"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.183310 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.183954 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.683932971 +0000 UTC m=+145.023581376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.259940 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.284491 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.285155 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.785143241 +0000 UTC m=+145.124791636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.285510 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.372415 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-l2x2j"] Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.387197 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.388612 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.888583397 +0000 UTC m=+145.228231802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.395141 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.395946 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.895929023 +0000 UTC m=+145.235577418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.432972 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-l55b8" podStartSLOduration=124.432954023 podStartE2EDuration="2m4.432954023s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.391112142 +0000 UTC m=+144.730760547" watchObservedRunningTime="2025-10-02 07:19:13.432954023 +0000 UTC m=+144.772602428" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.496259 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.496646 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:13.996632468 +0000 UTC m=+145.336280873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.546767 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-tjn2t" podStartSLOduration=124.546748334 podStartE2EDuration="2m4.546748334s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.509793147 +0000 UTC m=+144.849441572" watchObservedRunningTime="2025-10-02 07:19:13.546748334 +0000 UTC m=+144.886396739" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.547549 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-hwmtz" podStartSLOduration=124.547543628 podStartE2EDuration="2m4.547543628s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.545119166 +0000 UTC m=+144.884767571" watchObservedRunningTime="2025-10-02 07:19:13.547543628 +0000 UTC m=+144.887192033" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.551414 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cmbx8" event={"ID":"12133a89-e03b-4a3f-b665-c5d790315e98","Type":"ContainerStarted","Data":"1e8620c111e00d2505853c8cd7ce33339e8dea0553a9afe840627acd250b3ff0"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.561544 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" event={"ID":"6125f058-c403-48da-87a6-74db78012f40","Type":"ContainerStarted","Data":"aa796466a3d524b19b2f6726917972bc76b24da20f68c69aacb30b19d05e4770"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.596211 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" podStartSLOduration=123.5961848 podStartE2EDuration="2m3.5961848s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.593778729 +0000 UTC m=+144.933427134" watchObservedRunningTime="2025-10-02 07:19:13.5961848 +0000 UTC m=+144.935833205" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.598092 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.599427 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" event={"ID":"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2","Type":"ContainerStarted","Data":"06cbe01a688235cc264de23699b57a13557bfc09bc6ef21f6c9f38fa8173b9e5"} Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.599954 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.09993743 +0000 UTC m=+145.439585835 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.635534 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cmbx8" podStartSLOduration=124.635514898 podStartE2EDuration="2m4.635514898s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.634998273 +0000 UTC m=+144.974646678" watchObservedRunningTime="2025-10-02 07:19:13.635514898 +0000 UTC m=+144.975163303" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.641204 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" event={"ID":"4d955b93-e95e-4307-9993-07a7f60176bc","Type":"ContainerStarted","Data":"9eb532d0ada507714791b473239610fc3e9b801bff6d09acf7d857665b0291f8"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.643720 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" event={"ID":"472abadc-66db-4150-9054-9771d488ae44","Type":"ContainerStarted","Data":"4c63efa9fadb4c122b1cb4ae98c040223f845764100b14c97f499492c2f3d2b5"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.644928 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.654268 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" event={"ID":"e7b4750a-3c7e-4fa4-9534-40a895624522","Type":"ContainerStarted","Data":"5672517f39346c1bce9524cf2b1e494a713772cca4b50322cd4fc79ae69d460d"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.659065 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2tvsb" event={"ID":"4c7291f2-090f-4ab6-bcd5-203fee2bbce1","Type":"ContainerStarted","Data":"27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.659604 4829 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-rjdl6 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.659638 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" podUID="472abadc-66db-4150-9054-9771d488ae44" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.664288 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qbhhx" event={"ID":"4b0294ed-4a71-4a99-abf6-170b7ba35d7c","Type":"ContainerStarted","Data":"62bebc58901d84f822b3cf1c5b7bf4aca70c0c87ba35c10c9225348a736e2b97"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.684727 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" event={"ID":"3e3f0cd5-00ac-4ec9-bef4-69d2ec2d5517","Type":"ContainerStarted","Data":"a3b9581989be11a1469c6d7636cf816d41183966ad145eda0a87b27aaaf471db"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.698999 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" event={"ID":"2cc39185-906e-4b18-a359-b1fd011b74f4","Type":"ContainerStarted","Data":"1eb6e584347847a16983345379ebe34e391713a4c4d6634be8e6b879d084ee60"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.699856 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.701213 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.201194422 +0000 UTC m=+145.540842827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.702667 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ws4lp" event={"ID":"56c155fb-95c9-46fa-ad70-456dccb31776","Type":"ContainerStarted","Data":"d2f5c82a58b53bb59fe7767174d2a8fd02ca09adb0a468af57fd7cfb539809fe"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.703186 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.703649 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.203633944 +0000 UTC m=+145.543282349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.709455 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" event={"ID":"410e2990-0700-422a-aa64-e425907892c6","Type":"ContainerStarted","Data":"936c415df7a675106fa9048fa304045c9b612d04015009719b2ee1974ad74789"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.709527 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" event={"ID":"410e2990-0700-422a-aa64-e425907892c6","Type":"ContainerStarted","Data":"ed6fb457394bff6b6ce02e5994804ee8a74ed5c16b761a26535c4353184d7c02"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.711061 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" podStartSLOduration=124.711051993 podStartE2EDuration="2m4.711051993s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.677566126 +0000 UTC m=+145.017214531" watchObservedRunningTime="2025-10-02 07:19:13.711051993 +0000 UTC m=+145.050700398" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.711213 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2tvsb" podStartSLOduration=124.711209367 podStartE2EDuration="2m4.711209367s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.708705223 +0000 UTC m=+145.048353658" watchObservedRunningTime="2025-10-02 07:19:13.711209367 +0000 UTC m=+145.050857772" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.719033 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" event={"ID":"8706882f-e939-4ac9-b145-527cb52f9294","Type":"ContainerStarted","Data":"5dba925553d464183ca35d9f1ee8b6c6b6f43691259c2555c6ffdef1c5c4de91"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.721253 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" event={"ID":"ed93e809-cf53-450c-8993-5a69e97542bd","Type":"ContainerStarted","Data":"87b9df258b5effa2cddd8f9c6a8a9ea8e615861b0aa2601ac531e29ce5ababab"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.731754 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" event={"ID":"20cb3658-2a8d-49c1-8e8e-a7675d34833c","Type":"ContainerStarted","Data":"d1d006be05353da8284ee9bf29bef6ad8e85d47a3e2f07a8b84d428198732fb4"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.740119 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-vm74c" podStartSLOduration=124.740101608 podStartE2EDuration="2m4.740101608s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.738559782 +0000 UTC m=+145.078208177" watchObservedRunningTime="2025-10-02 07:19:13.740101608 +0000 UTC m=+145.079750013" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.772620 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" event={"ID":"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a","Type":"ContainerStarted","Data":"30c219a07ea7a3e8f0dd4125c1dd8f2a58047f3194b0c451a53d8875cd1eee71"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.772652 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" event={"ID":"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a","Type":"ContainerStarted","Data":"4056463d6bd304ccbd92d9ea70201ef90ae709ecd63d7fa6ee5f6469510a64ba"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.806022 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.809824 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.30980015 +0000 UTC m=+145.649448615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.852388 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" event={"ID":"835adc97-13f4-4100-811b-72d0cc4a2ed3","Type":"ContainerStarted","Data":"9675948a961bd46966f5296c00bcb08eeefa24e789cd334909b7eff5717dbaf9"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.852448 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" event={"ID":"835adc97-13f4-4100-811b-72d0cc4a2ed3","Type":"ContainerStarted","Data":"30f8f7fc4a871c0aa016059ad93b634beb4ee44ce37f21c355ff7e176df09c9b"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.860376 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-mcxkp" podStartSLOduration=123.860358719 podStartE2EDuration="2m3.860358719s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.782951309 +0000 UTC m=+145.122599714" watchObservedRunningTime="2025-10-02 07:19:13.860358719 +0000 UTC m=+145.200007124" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.862119 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" event={"ID":"e639b9b5-75b2-43af-a51c-9402b7dcec03","Type":"ContainerStarted","Data":"0bd83e5354010b3aafe82d29d90dc4585b6bfa4f2558c87248a9d4ac9b795f97"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.871706 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-h2g87" podStartSLOduration=124.871693793 podStartE2EDuration="2m4.871693793s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.831597062 +0000 UTC m=+145.171245467" watchObservedRunningTime="2025-10-02 07:19:13.871693793 +0000 UTC m=+145.211342198" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.872862 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-br5pp" podStartSLOduration=124.872857697 podStartE2EDuration="2m4.872857697s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.869833128 +0000 UTC m=+145.209481533" watchObservedRunningTime="2025-10-02 07:19:13.872857697 +0000 UTC m=+145.212506102" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.891105 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" event={"ID":"2e0e4181-e7fd-42fc-90b8-5e145f641943","Type":"ContainerStarted","Data":"ac66a22531034d0affa1529759b983ac2f8598f419b9bc97a55b6b616367e33c"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.891704 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.895451 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" event={"ID":"900dfada-c31e-41ec-97a4-2244cf0d8781","Type":"ContainerStarted","Data":"8bd752efa0a8523e4933761c879840725df326a0fdfa4216352f3e0eeb24ee90"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.902557 4829 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6w49n container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" start-of-body= Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.902626 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": dial tcp 10.217.0.32:6443: connect: connection refused" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.919858 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:13 crc kubenswrapper[4829]: E1002 07:19:13.920408 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.420394966 +0000 UTC m=+145.760043371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.932169 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" event={"ID":"d0e365b9-9ffa-412b-8055-559480d9d565","Type":"ContainerStarted","Data":"9c43eaab931c350998bd33c55aa95c5361c78c76a6fa4f7eb440437d04c3e77e"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.934354 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" event={"ID":"129ca2ba-597b-40f3-a871-d7196b84b4f0","Type":"ContainerStarted","Data":"2d2843c0008589d83dd2e640a10f56c6fbccb6e96e2c55f09c5d49e0bb680cc5"} Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.951300 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-8nnbv" podStartSLOduration=124.951278316 podStartE2EDuration="2m4.951278316s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.949834383 +0000 UTC m=+145.289482788" watchObservedRunningTime="2025-10-02 07:19:13.951278316 +0000 UTC m=+145.290926711" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.952306 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" podStartSLOduration=124.952300296 podStartE2EDuration="2m4.952300296s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.919947923 +0000 UTC m=+145.259596328" watchObservedRunningTime="2025-10-02 07:19:13.952300296 +0000 UTC m=+145.291948701" Oct 02 07:19:13 crc kubenswrapper[4829]: I1002 07:19:13.981421 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" event={"ID":"96e0038e-e6bd-497c-8245-8f11f8428a00","Type":"ContainerStarted","Data":"7eea9d795dbf1c1b700aebccfe617fc40ff2d4d70f78ac28753692696236724a"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:13.995040 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" event={"ID":"d5eef982-bef6-4679-be23-88feef0e3793","Type":"ContainerStarted","Data":"d9ab2e6df7d6cdace05d035f53da7c774e06c67d7f350b50191002ee0086f8c0"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:13.998166 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" event={"ID":"5443f78e-7898-4ee6-9f8f-574331118acd","Type":"ContainerStarted","Data":"1fdb74c4020626a7f38dcd5e7ff53acb454ca7c0886e2476f86b387e8ca15c1b"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.010485 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" podStartSLOduration=125.010436948 podStartE2EDuration="2m5.010436948s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:13.999339571 +0000 UTC m=+145.338987976" watchObservedRunningTime="2025-10-02 07:19:14.010436948 +0000 UTC m=+145.350085353" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.020517 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-xbjjp" event={"ID":"dddb80c0-70db-4c01-8bfb-ce39a6cd1406","Type":"ContainerStarted","Data":"27473430ba8b7ab16241652971bfbf14e2187c687376fdd632f1a60a14a14d05"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.021311 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.023661 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.523632457 +0000 UTC m=+145.863280862 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.041265 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" event={"ID":"84c837e4-1328-4abd-ac80-d75637b192d1","Type":"ContainerStarted","Data":"9d8bb104f49ddfffeea507be1351278df5b592f26901b12d948975e129458139"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.052060 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.052110 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" event={"ID":"84c837e4-1328-4abd-ac80-d75637b192d1","Type":"ContainerStarted","Data":"30644c8f0a0229fe849aaaf6825fc81b69994cfcceacc38436be2911b518c3d7"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.052737 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" event={"ID":"40922895-36ea-484e-a030-2e9ea196629e","Type":"ContainerStarted","Data":"1314c8692acf207eea91e375ed873389a812fd69a55179496d5746b8be813bc2"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.055486 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" event={"ID":"5a36f298-4789-4e89-91a3-d7baef3ed8b2","Type":"ContainerStarted","Data":"8bec740fe53fbd24de11f18bee27f56926c4806e1f4cc6a63d9902651e75f866"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.055511 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" event={"ID":"5a36f298-4789-4e89-91a3-d7baef3ed8b2","Type":"ContainerStarted","Data":"2009403c0f61f5d1d8253051442045ec1f842a2b20057e09f7bf6c00046a4e1d"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.063013 4829 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-4dk4r container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.063069 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" podUID="84c837e4-1328-4abd-ac80-d75637b192d1" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.076623 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-49p7f" podStartSLOduration=125.076602556 podStartE2EDuration="2m5.076602556s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.050605561 +0000 UTC m=+145.390253966" watchObservedRunningTime="2025-10-02 07:19:14.076602556 +0000 UTC m=+145.416250961" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.102101 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" event={"ID":"100e06e7-e40d-4b3f-81cf-86886ad1e928","Type":"ContainerStarted","Data":"32b13c0e4ca98e4d0bdc3eb8e19f5db513450c08cc500bbad27bda6e4e7cbe21"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.124355 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.126822 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" event={"ID":"34bc2780-c7b8-45cf-9924-4b08def7fd64","Type":"ContainerStarted","Data":"c77c023b43bdfbe72b1ca97fefc1123f63f21c82323976f418b73d5328fd5556"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.127413 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.128008 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.62799596 +0000 UTC m=+145.967644365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.134918 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" podStartSLOduration=125.134901423 podStartE2EDuration="2m5.134901423s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.127684671 +0000 UTC m=+145.467333066" watchObservedRunningTime="2025-10-02 07:19:14.134901423 +0000 UTC m=+145.474549828" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.148959 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" event={"ID":"4481fcdc-864e-4b94-8378-78cb9241dd12","Type":"ContainerStarted","Data":"13da5ad920745f5a116b927b366cd3a00517e171af0951106a2e8183153b7d0d"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.150503 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" event={"ID":"fcb6162b-fb97-4bf5-8563-48c907aa6a36","Type":"ContainerStarted","Data":"b2798d12b7e52da1b87aa190e869298164544f4c04dd37df37da2ac38a8041dd"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.154464 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-r7cxf" event={"ID":"ca598dbb-d00b-4361-88c0-34c22ad2be6e","Type":"ContainerStarted","Data":"c946b77eafbd01d14650a24c7280b374e3141bcf99ece2bd6d7358936e1d697f"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.160417 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.164705 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kcs46" podStartSLOduration=125.16468986 podStartE2EDuration="2m5.16468986s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.160365503 +0000 UTC m=+145.500013898" watchObservedRunningTime="2025-10-02 07:19:14.16468986 +0000 UTC m=+145.504338255" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.170143 4829 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7cxf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.170391 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7cxf" podUID="ca598dbb-d00b-4361-88c0-34c22ad2be6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.173992 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-26l22" event={"ID":"b1f5af08-068d-43b7-aaa5-45da6c35918c","Type":"ContainerStarted","Data":"622b0da950842ff9bd0032eace76d69d80cbd94695abb46c4fee3c56bec9fe89"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.179242 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" event={"ID":"72e0891e-a4b5-43c2-918f-e5ffdc53fe20","Type":"ContainerStarted","Data":"fda293488888519a85d3094cd9573cc5bc8dd047d59a3e8db25b46a428d7e9d2"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.183734 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rnq88" event={"ID":"7a9ac0f1-deb2-4be2-9aa2-a2bf69c7d979","Type":"ContainerStarted","Data":"b643ca083714d2c0b5e55e46dce40565e371609ef322a5e53a57922a8166bcb8"} Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.197370 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9v6qs" podStartSLOduration=125.197354102 podStartE2EDuration="2m5.197354102s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.197156727 +0000 UTC m=+145.536805132" watchObservedRunningTime="2025-10-02 07:19:14.197354102 +0000 UTC m=+145.537002507" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.226388 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.228204 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.72818768 +0000 UTC m=+146.067836085 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.234702 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" podStartSLOduration=125.234684021 podStartE2EDuration="2m5.234684021s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.222787661 +0000 UTC m=+145.562436066" watchObservedRunningTime="2025-10-02 07:19:14.234684021 +0000 UTC m=+145.574332426" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.244720 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" podStartSLOduration=125.244707007 podStartE2EDuration="2m5.244707007s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.241695918 +0000 UTC m=+145.581344323" watchObservedRunningTime="2025-10-02 07:19:14.244707007 +0000 UTC m=+145.584355412" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.283318 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" podStartSLOduration=124.283298643 podStartE2EDuration="2m4.283298643s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.282945832 +0000 UTC m=+145.622594237" watchObservedRunningTime="2025-10-02 07:19:14.283298643 +0000 UTC m=+145.622947048" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.335372 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.335658 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.835648354 +0000 UTC m=+146.175296759 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.347908 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6fsbb" podStartSLOduration=124.347891085 podStartE2EDuration="2m4.347891085s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.305054013 +0000 UTC m=+145.644702418" watchObservedRunningTime="2025-10-02 07:19:14.347891085 +0000 UTC m=+145.687539490" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.350055 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-xbjjp" podStartSLOduration=6.350021868 podStartE2EDuration="6.350021868s" podCreationTimestamp="2025-10-02 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.34536444 +0000 UTC m=+145.685012845" watchObservedRunningTime="2025-10-02 07:19:14.350021868 +0000 UTC m=+145.689670273" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.384950 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-26l22" podStartSLOduration=125.384923665 podStartE2EDuration="2m5.384923665s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.381955008 +0000 UTC m=+145.721603413" watchObservedRunningTime="2025-10-02 07:19:14.384923665 +0000 UTC m=+145.724572070" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.385263 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.409841 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:14 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:14 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:14 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.410476 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.436317 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.436709 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:14.93669089 +0000 UTC m=+146.276339295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.475354 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-r7cxf" podStartSLOduration=125.475335337 podStartE2EDuration="2m5.475335337s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.439027618 +0000 UTC m=+145.778676023" watchObservedRunningTime="2025-10-02 07:19:14.475335337 +0000 UTC m=+145.814983742" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.475888 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rnq88" podStartSLOduration=125.475884073 podStartE2EDuration="2m5.475884073s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:14.472848794 +0000 UTC m=+145.812497209" watchObservedRunningTime="2025-10-02 07:19:14.475884073 +0000 UTC m=+145.815532478" Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.538685 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.539020 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.039008732 +0000 UTC m=+146.378657137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.639538 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.640127 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.140111119 +0000 UTC m=+146.479759524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.742010 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.742336 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.242324419 +0000 UTC m=+146.581972824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.843251 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.843420 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.343393215 +0000 UTC m=+146.683041620 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.843485 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.843798 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.343791317 +0000 UTC m=+146.683439722 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.944358 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.944563 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.444540603 +0000 UTC m=+146.784189008 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:14 crc kubenswrapper[4829]: I1002 07:19:14.944905 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:14 crc kubenswrapper[4829]: E1002 07:19:14.945301 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.445289005 +0000 UTC m=+146.784937410 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.045770 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.046193 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.546178527 +0000 UTC m=+146.885826932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.147904 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.148307 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.648289393 +0000 UTC m=+146.987937798 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.187734 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-ws4lp" event={"ID":"56c155fb-95c9-46fa-ad70-456dccb31776","Type":"ContainerStarted","Data":"c54451cc7160e801a907091f91c1c63c8c8e96b215da5793e8024d350d578804"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.189130 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" event={"ID":"4481fcdc-864e-4b94-8378-78cb9241dd12","Type":"ContainerStarted","Data":"31e69cc2be0558ca8364c5edab4c08ddf60ad13dc4867c5514010302d27a0fa3"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.189160 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" event={"ID":"4481fcdc-864e-4b94-8378-78cb9241dd12","Type":"ContainerStarted","Data":"29b9bcfcb9442ec343c70bf2ecb0906cf39c6d25379814b0819d739b17d08be9"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.191047 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" event={"ID":"d0e365b9-9ffa-412b-8055-559480d9d565","Type":"ContainerStarted","Data":"3c14e14ea23ebf0a4e286124d461415d692edfa24789cfcd7978ff47c96b242c"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.192846 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-26l22" event={"ID":"b1f5af08-068d-43b7-aaa5-45da6c35918c","Type":"ContainerStarted","Data":"e48bfb3168608563a675cd9dfdbedca0fdcef1a473e1eaf7303dc9a3b556cf7d"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.195110 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" event={"ID":"835adc97-13f4-4100-811b-72d0cc4a2ed3","Type":"ContainerStarted","Data":"04dc06ac0244e20f44681d694a9c8bd85fa3eb21f9029a795ba14e7f960981ed"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.197016 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qbhhx" event={"ID":"4b0294ed-4a71-4a99-abf6-170b7ba35d7c","Type":"ContainerStarted","Data":"88151164c28051371592c9ffe283d1594dbc31a972706b1d7aa6da0d877c5e03"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.197040 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-qbhhx" event={"ID":"4b0294ed-4a71-4a99-abf6-170b7ba35d7c","Type":"ContainerStarted","Data":"711e8c1d862a99d64a1738720a9aeaef3369fbd9a278e7f0333f6c70c42a9d70"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.197392 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.198554 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" event={"ID":"fcb6162b-fb97-4bf5-8563-48c907aa6a36","Type":"ContainerStarted","Data":"81c3317d96541fc3d32359783889df7c22cd2635a4bd5f4767bfbebcc455e0b1"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.199659 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" event={"ID":"129ca2ba-597b-40f3-a871-d7196b84b4f0","Type":"ContainerStarted","Data":"7f6b048d5d1525993733cf36db8d1d3bc71d906b9fbabaffc26a63ffecb3ad1e"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.200293 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.201487 4829 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ghvdx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" start-of-body= Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.201561 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" podUID="129ca2ba-597b-40f3-a871-d7196b84b4f0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": dial tcp 10.217.0.38:5443: connect: connection refused" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.201590 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" event={"ID":"4d955b93-e95e-4307-9993-07a7f60176bc","Type":"ContainerStarted","Data":"572fd150d0a502c61fbb694fe0331ed6f9aefbca59a0b86fff6b5aff5845fab7"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.203166 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" event={"ID":"8706882f-e939-4ac9-b145-527cb52f9294","Type":"ContainerStarted","Data":"3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.203807 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.204479 4829 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tsrrk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.204513 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" podUID="8706882f-e939-4ac9-b145-527cb52f9294" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.205164 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" event={"ID":"40922895-36ea-484e-a030-2e9ea196629e","Type":"ContainerStarted","Data":"64b0b8c7b7de544c359042bfe9be5cd23d0ad0cbdf66f241afc4ac4828314e57"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.205194 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" event={"ID":"40922895-36ea-484e-a030-2e9ea196629e","Type":"ContainerStarted","Data":"a64320e56db14375cd1484292070412c1f58d033d1d71d13057edd78c386984a"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.206696 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zqd5d" event={"ID":"2cc39185-906e-4b18-a359-b1fd011b74f4","Type":"ContainerStarted","Data":"4aa54a1710fb6a80d48ba1301c8055ec76dd8f6987529b5be2e6d0886a295edb"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.208950 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" event={"ID":"72e0891e-a4b5-43c2-918f-e5ffdc53fe20","Type":"ContainerStarted","Data":"8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.209597 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.210448 4829 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r77tn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.210481 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.211127 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" event={"ID":"6125f058-c403-48da-87a6-74db78012f40","Type":"ContainerStarted","Data":"0f03729d0208caa874a32a25d517598d3903c09e8fbd10e331d1f55ac5f1d2b0"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.211152 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" event={"ID":"6125f058-c403-48da-87a6-74db78012f40","Type":"ContainerStarted","Data":"dc6330821413fb625f06167c610e8d22796b60021966e74e801f6ceea2bd15af"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.212930 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" event={"ID":"e639b9b5-75b2-43af-a51c-9402b7dcec03","Type":"ContainerStarted","Data":"5599ca9df1def466e998b03cd3f5c5466906d41841eb089ad051f37edf3a855a"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.212955 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" event={"ID":"e639b9b5-75b2-43af-a51c-9402b7dcec03","Type":"ContainerStarted","Data":"17c2e33ed4fae7f913beda9c015c2062f07a24b18e5d63c7901a91d9ed39f16c"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.213291 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.217433 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" event={"ID":"72cf0023-3aa0-4ee6-9696-fb3155c8f8c2","Type":"ContainerStarted","Data":"938867d6c37da41958cfdd8d7b2c2e9621ddc2ca2bae456b00995094b23b09c3"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.219472 4829 generic.go:334] "Generic (PLEG): container finished" podID="5443f78e-7898-4ee6-9f8f-574331118acd" containerID="8de398c8c50cab14edc127b36e2d085a1804dd44e848951858e7761910df7dc0" exitCode=0 Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.220273 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" event={"ID":"5443f78e-7898-4ee6-9f8f-574331118acd","Type":"ContainerDied","Data":"8de398c8c50cab14edc127b36e2d085a1804dd44e848951858e7761910df7dc0"} Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.223287 4829 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7cxf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.223333 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7cxf" podUID="ca598dbb-d00b-4361-88c0-34c22ad2be6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.224020 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.237964 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-rjdl6" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.243300 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-4dk4r" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.252923 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.254098 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.754081968 +0000 UTC m=+147.093730373 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.254400 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.254769 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.754752418 +0000 UTC m=+147.094400823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.263326 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-ws4lp" podStartSLOduration=7.26331025 podStartE2EDuration="7.26331025s" podCreationTimestamp="2025-10-02 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.217564174 +0000 UTC m=+146.557212579" watchObservedRunningTime="2025-10-02 07:19:15.26331025 +0000 UTC m=+146.602958655" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.263659 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ctfz2" podStartSLOduration=126.263652741 podStartE2EDuration="2m6.263652741s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.262426444 +0000 UTC m=+146.602074849" watchObservedRunningTime="2025-10-02 07:19:15.263652741 +0000 UTC m=+146.603301156" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.304717 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-mpm4h" podStartSLOduration=126.304704269 podStartE2EDuration="2m6.304704269s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.303123232 +0000 UTC m=+146.642771637" watchObservedRunningTime="2025-10-02 07:19:15.304704269 +0000 UTC m=+146.644352674" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.360667 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.363068 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.863053177 +0000 UTC m=+147.202701582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.390195 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:15 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:15 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:15 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.390503 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.409030 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" podStartSLOduration=126.409015471 podStartE2EDuration="2m6.409015471s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.40866861 +0000 UTC m=+146.748317015" watchObservedRunningTime="2025-10-02 07:19:15.409015471 +0000 UTC m=+146.748663876" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.436730 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rnq88" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.448084 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-dk9k6" podStartSLOduration=126.44806785 podStartE2EDuration="2m6.44806785s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.447578486 +0000 UTC m=+146.787226891" watchObservedRunningTime="2025-10-02 07:19:15.44806785 +0000 UTC m=+146.787716255" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.464373 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.464809 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:15.964796013 +0000 UTC m=+147.304444418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.521037 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-f68cm" podStartSLOduration=126.521017678 podStartE2EDuration="2m6.521017678s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.494540309 +0000 UTC m=+146.834188724" watchObservedRunningTime="2025-10-02 07:19:15.521017678 +0000 UTC m=+146.860666083" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.565038 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.565423 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.065408446 +0000 UTC m=+147.405056851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.593080 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" podStartSLOduration=126.59306387 podStartE2EDuration="2m6.59306387s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.590957968 +0000 UTC m=+146.930606383" watchObservedRunningTime="2025-10-02 07:19:15.59306387 +0000 UTC m=+146.932712275" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.663823 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-qbhhx" podStartSLOduration=7.663809003 podStartE2EDuration="7.663809003s" podCreationTimestamp="2025-10-02 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.660724172 +0000 UTC m=+147.000372587" watchObservedRunningTime="2025-10-02 07:19:15.663809003 +0000 UTC m=+147.003457408" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.670550 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.670953 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.170935033 +0000 UTC m=+147.510583438 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.772308 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.772611 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.272597076 +0000 UTC m=+147.612245481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.809602 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" podStartSLOduration=126.809583036 podStartE2EDuration="2m6.809583036s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.729535368 +0000 UTC m=+147.069183773" watchObservedRunningTime="2025-10-02 07:19:15.809583036 +0000 UTC m=+147.149231441" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.811156 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" podStartSLOduration=125.811148832 podStartE2EDuration="2m5.811148832s" podCreationTimestamp="2025-10-02 07:17:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.80801651 +0000 UTC m=+147.147664905" watchObservedRunningTime="2025-10-02 07:19:15.811148832 +0000 UTC m=+147.150797237" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.873274 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.873625 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.373614681 +0000 UTC m=+147.713263076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.874556 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-4dgbl" podStartSLOduration=126.874539408 podStartE2EDuration="2m6.874539408s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.872639993 +0000 UTC m=+147.212288398" watchObservedRunningTime="2025-10-02 07:19:15.874539408 +0000 UTC m=+147.214187803" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.874791 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kzcst" podStartSLOduration=126.874786445 podStartE2EDuration="2m6.874786445s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.844271557 +0000 UTC m=+147.183919962" watchObservedRunningTime="2025-10-02 07:19:15.874786445 +0000 UTC m=+147.214434850" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.937528 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6jt2q" podStartSLOduration=126.937511493 podStartE2EDuration="2m6.937511493s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:15.904292824 +0000 UTC m=+147.243941229" watchObservedRunningTime="2025-10-02 07:19:15.937511493 +0000 UTC m=+147.277159888" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.966703 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.966781 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.974125 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.974251 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.474212433 +0000 UTC m=+147.813860838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:15 crc kubenswrapper[4829]: I1002 07:19:15.974379 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:15 crc kubenswrapper[4829]: E1002 07:19:15.974807 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.474771679 +0000 UTC m=+147.814420084 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.077653 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.078195 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.578181625 +0000 UTC m=+147.917830030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.179109 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.179533 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.679517658 +0000 UTC m=+148.019166063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.224096 4829 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6w49n container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.32:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.224163 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.32:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.227473 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" event={"ID":"5443f78e-7898-4ee6-9f8f-574331118acd","Type":"ContainerStarted","Data":"d177e07c982990c37113f2d4b348caaef5eefbce833932dc9b31f26a348f14f6"} Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.227941 4829 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r77tn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.227981 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.239820 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.249013 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-226j2" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.282537 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.282868 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.782853302 +0000 UTC m=+148.122501707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.378920 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.384478 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.385651 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.885638748 +0000 UTC m=+148.225287153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.394199 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:16 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:16 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:16 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.394301 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.402194 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.488456 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.488838 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:16.988821836 +0000 UTC m=+148.328470241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.590086 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.590542 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.090524661 +0000 UTC m=+148.430173066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.692798 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.693097 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.19308267 +0000 UTC m=+148.532731075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.794329 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.794630 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.294618661 +0000 UTC m=+148.634267066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.895489 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.895672 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.395647025 +0000 UTC m=+148.735295430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.895776 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.896040 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.396026817 +0000 UTC m=+148.735675222 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.997053 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.997194 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.497177245 +0000 UTC m=+148.836825650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:16 crc kubenswrapper[4829]: I1002 07:19:16.997325 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:16 crc kubenswrapper[4829]: E1002 07:19:16.997585 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.497577337 +0000 UTC m=+148.837225742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.072526 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7rgt2"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.073394 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.090706 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.097825 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.097977 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.597956543 +0000 UTC m=+148.937604948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.098331 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.098615 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.598606492 +0000 UTC m=+148.938254977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.136289 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7rgt2"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.199512 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.199679 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.699650807 +0000 UTC m=+149.039299212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.199775 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.199819 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-utilities\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.199879 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb9kg\" (UniqueName: \"kubernetes.io/projected/0357590d-1dad-4ef3-b422-8b25a71953ab-kube-api-access-gb9kg\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.199905 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-catalog-content\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.200111 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.70010369 +0000 UTC m=+149.039752095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.228465 4829 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ghvdx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.228730 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" podUID="129ca2ba-597b-40f3-a871-d7196b84b4f0" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.38:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.234099 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" event={"ID":"5443f78e-7898-4ee6-9f8f-574331118acd","Type":"ContainerStarted","Data":"9a288b672f68c74a7af2b61efb7ff11364436f9e23fa2e1c687deb25abcd51d3"} Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.236205 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" event={"ID":"fcb6162b-fb97-4bf5-8563-48c907aa6a36","Type":"ContainerStarted","Data":"61e9edf21f093587e387344127ebc7d7566be9ec09677ba184ac7b1a8b8f03e8"} Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.236262 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" event={"ID":"fcb6162b-fb97-4bf5-8563-48c907aa6a36","Type":"ContainerStarted","Data":"ab8e265650c476f65ad67fc73d79a50d25c9e6f5cd904e62e25681859acc6f4c"} Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.236273 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" event={"ID":"fcb6162b-fb97-4bf5-8563-48c907aa6a36","Type":"ContainerStarted","Data":"cf0e077dda2f851a7222f52131c17890e94377034e3b6c322fe6f581bf18ba99"} Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.236725 4829 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-r77tn container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" start-of-body= Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.236784 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": dial tcp 10.217.0.35:8080: connect: connection refused" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.242510 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wfrrj" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.259113 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" podStartSLOduration=128.259098547 podStartE2EDuration="2m8.259098547s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:17.256062058 +0000 UTC m=+148.595710483" watchObservedRunningTime="2025-10-02 07:19:17.259098547 +0000 UTC m=+148.598746952" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.279096 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xl9qq"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.279984 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.298774 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.302154 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.302597 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb9kg\" (UniqueName: \"kubernetes.io/projected/0357590d-1dad-4ef3-b422-8b25a71953ab-kube-api-access-gb9kg\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.302657 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-catalog-content\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.302780 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.302934 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-utilities\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.302956 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.303293 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xl9qq"] Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.303472 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.803453423 +0000 UTC m=+149.143101828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.306149 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-utilities\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.306831 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-catalog-content\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.308305 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ghvdx" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.313708 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.321020 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.346025 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb9kg\" (UniqueName: \"kubernetes.io/projected/0357590d-1dad-4ef3-b422-8b25a71953ab-kube-api-access-gb9kg\") pod \"community-operators-7rgt2\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.364507 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-psj8g" podStartSLOduration=9.364492691 podStartE2EDuration="9.364492691s" podCreationTimestamp="2025-10-02 07:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:17.362275046 +0000 UTC m=+148.701923461" watchObservedRunningTime="2025-10-02 07:19:17.364492691 +0000 UTC m=+148.704141096" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.385050 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.385935 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:17 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:17 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:17 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.386001 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.406860 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.406903 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.406965 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnq4t\" (UniqueName: \"kubernetes.io/projected/3cd1d972-be99-4ae8-b965-dd55a308e3a9-kube-api-access-fnq4t\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.407008 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-utilities\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.407028 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.407052 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-catalog-content\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.407332 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:17.907320972 +0000 UTC m=+149.246969377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.411293 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.420709 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.482056 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.484746 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rltqp"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.486188 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.490992 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.491807 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rltqp"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.497864 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.508559 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.508942 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-utilities\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.508978 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-catalog-content\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.509047 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnq4t\" (UniqueName: \"kubernetes.io/projected/3cd1d972-be99-4ae8-b965-dd55a308e3a9-kube-api-access-fnq4t\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.509423 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.009408568 +0000 UTC m=+149.349056973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.509760 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-utilities\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.509955 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-catalog-content\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.564053 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnq4t\" (UniqueName: \"kubernetes.io/projected/3cd1d972-be99-4ae8-b965-dd55a308e3a9-kube-api-access-fnq4t\") pod \"certified-operators-xl9qq\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.592514 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.605330 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.605625 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.611448 4829 patch_prober.go:28] interesting pod/apiserver-76f77b778f-l2x2j container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.611528 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" podUID="5443f78e-7898-4ee6-9f8f-574331118acd" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.614850 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-utilities\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.614888 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-catalog-content\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.614924 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.614970 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77k7x\" (UniqueName: \"kubernetes.io/projected/330c4c1e-cb32-42fd-a3a8-8256339e5e17-kube-api-access-77k7x\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.615277 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.115265945 +0000 UTC m=+149.454914350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.651426 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4lkc7"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.654277 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.671160 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4lkc7"] Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.681551 4829 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724243 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724440 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhd7n\" (UniqueName: \"kubernetes.io/projected/8d3333e7-33d0-4f71-b051-6980f83a77a4-kube-api-access-dhd7n\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724468 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-catalog-content\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724501 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77k7x\" (UniqueName: \"kubernetes.io/projected/330c4c1e-cb32-42fd-a3a8-8256339e5e17-kube-api-access-77k7x\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724550 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-utilities\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724568 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-catalog-content\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.724587 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-utilities\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.724683 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.224667836 +0000 UTC m=+149.564316241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.725296 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-utilities\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.725490 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-catalog-content\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.758757 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77k7x\" (UniqueName: \"kubernetes.io/projected/330c4c1e-cb32-42fd-a3a8-8256339e5e17-kube-api-access-77k7x\") pod \"community-operators-rltqp\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.825602 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-catalog-content\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.825722 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-utilities\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.825747 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.825794 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhd7n\" (UniqueName: \"kubernetes.io/projected/8d3333e7-33d0-4f71-b051-6980f83a77a4-kube-api-access-dhd7n\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.826565 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-catalog-content\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.826829 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-utilities\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.827090 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.327080473 +0000 UTC m=+149.666728878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.842599 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.885964 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhd7n\" (UniqueName: \"kubernetes.io/projected/8d3333e7-33d0-4f71-b051-6980f83a77a4-kube-api-access-dhd7n\") pod \"certified-operators-4lkc7\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:17 crc kubenswrapper[4829]: I1002 07:19:17.927074 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:17 crc kubenswrapper[4829]: E1002 07:19:17.927485 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.427470558 +0000 UTC m=+149.767118963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.029859 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.030208 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.530188163 +0000 UTC m=+149.869836568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.053502 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.132828 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.133293 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.633274639 +0000 UTC m=+149.972923054 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.206776 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7rgt2"] Oct 02 07:19:18 crc kubenswrapper[4829]: W1002 07:19:18.236349 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0357590d_1dad_4ef3_b422_8b25a71953ab.slice/crio-5bf331f22acbd276e8317f5908689df5c15159c34ce5285e336ae709584d232c WatchSource:0}: Error finding container 5bf331f22acbd276e8317f5908689df5c15159c34ce5285e336ae709584d232c: Status 404 returned error can't find the container with id 5bf331f22acbd276e8317f5908689df5c15159c34ce5285e336ae709584d232c Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.237023 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.237334 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.737322372 +0000 UTC m=+150.076970777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.298067 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"77ef4d80745dd7506442cc6ba07a0e42a8dce2c93355f664f7b5b57e1e9a71c6"} Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.338754 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.339091 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.839076758 +0000 UTC m=+150.178725163 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.356746 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xl9qq"] Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.387336 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:18 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:18 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:18 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.387384 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.443984 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.446603 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:18.946591624 +0000 UTC m=+150.286240029 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.544701 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.545104 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 07:19:19.045089224 +0000 UTC m=+150.384737629 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.631723 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4lkc7"] Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.647889 4829 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T07:19:17.681576218Z","Handler":null,"Name":""} Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.648142 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:18 crc kubenswrapper[4829]: E1002 07:19:18.648579 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 07:19:19.148566082 +0000 UTC m=+150.488214477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qx6xb" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.657084 4829 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.657114 4829 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.691455 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rltqp"] Oct 02 07:19:18 crc kubenswrapper[4829]: W1002 07:19:18.718827 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod330c4c1e_cb32_42fd_a3a8_8256339e5e17.slice/crio-a2285a7615f4768f6fb3ba1121494581344af3fa8ebccaf5aad0581967f4cb6c WatchSource:0}: Error finding container a2285a7615f4768f6fb3ba1121494581344af3fa8ebccaf5aad0581967f4cb6c: Status 404 returned error can't find the container with id a2285a7615f4768f6fb3ba1121494581344af3fa8ebccaf5aad0581967f4cb6c Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.749181 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.780760 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.851919 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.865331 4829 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.865378 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:18 crc kubenswrapper[4829]: I1002 07:19:18.950851 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qx6xb\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.098849 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.245509 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7cm84"] Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.248130 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.252140 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.255497 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cm84"] Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.319269 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"50a1b289e39b3b0e0e853850dd569bb58b7e2a5a67f5b33ffec673f98abf08ed"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.350045 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d6a587f43ccd7b4abdee1db01daec3fd14f857b7aed1f8866ff94b1e2f1b7ada"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.350093 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6e9bb290772e452715596457c95a247c0c6de9afb0f2c985230ffb2723d2a522"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.355040 4829 generic.go:334] "Generic (PLEG): container finished" podID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerID="b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3" exitCode=0 Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.355102 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lkc7" event={"ID":"8d3333e7-33d0-4f71-b051-6980f83a77a4","Type":"ContainerDied","Data":"b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.355126 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lkc7" event={"ID":"8d3333e7-33d0-4f71-b051-6980f83a77a4","Type":"ContainerStarted","Data":"995ebc9b1c60ae1b74c8ed0aff80dad26609f95ca5453903906db8d16ea535e1"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.358064 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.365501 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-utilities\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.365730 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-catalog-content\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.365788 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv4s2\" (UniqueName: \"kubernetes.io/projected/37c0917f-a686-4dc9-99d3-28ff284ca2e0-kube-api-access-fv4s2\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.368390 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f5f95afaedef8e595f4896cb69699b4cf858a01b500b61d244e6a2b6a29b648e"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.368441 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"29dcea14fce53468bd18626290e766f33747329c52b52d99a4cd984f1bc86f92"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.369063 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.371022 4829 generic.go:334] "Generic (PLEG): container finished" podID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerID="ae33fa9287d69a7524a09ec08cc9a5d39c61ec3fb2fa7d4936cb372c26499f70" exitCode=0 Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.371063 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl9qq" event={"ID":"3cd1d972-be99-4ae8-b965-dd55a308e3a9","Type":"ContainerDied","Data":"ae33fa9287d69a7524a09ec08cc9a5d39c61ec3fb2fa7d4936cb372c26499f70"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.371078 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl9qq" event={"ID":"3cd1d972-be99-4ae8-b965-dd55a308e3a9","Type":"ContainerStarted","Data":"64b576ad586a8f4e030892b0b0502accd8cad58593ecbf288ed9a40fbf3b338f"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.388684 4829 generic.go:334] "Generic (PLEG): container finished" podID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerID="54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c" exitCode=0 Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.388747 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rgt2" event={"ID":"0357590d-1dad-4ef3-b422-8b25a71953ab","Type":"ContainerDied","Data":"54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.388774 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rgt2" event={"ID":"0357590d-1dad-4ef3-b422-8b25a71953ab","Type":"ContainerStarted","Data":"5bf331f22acbd276e8317f5908689df5c15159c34ce5285e336ae709584d232c"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.391297 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:19 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:19 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:19 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.391330 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.409806 4829 generic.go:334] "Generic (PLEG): container finished" podID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerID="0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82" exitCode=0 Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.409943 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rltqp" event={"ID":"330c4c1e-cb32-42fd-a3a8-8256339e5e17","Type":"ContainerDied","Data":"0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.410028 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rltqp" event={"ID":"330c4c1e-cb32-42fd-a3a8-8256339e5e17","Type":"ContainerStarted","Data":"a2285a7615f4768f6fb3ba1121494581344af3fa8ebccaf5aad0581967f4cb6c"} Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.455812 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qx6xb"] Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.466851 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-catalog-content\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.466897 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv4s2\" (UniqueName: \"kubernetes.io/projected/37c0917f-a686-4dc9-99d3-28ff284ca2e0-kube-api-access-fv4s2\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.466995 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-utilities\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.472294 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.475511 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-utilities\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.476332 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-catalog-content\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.500126 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv4s2\" (UniqueName: \"kubernetes.io/projected/37c0917f-a686-4dc9-99d3-28ff284ca2e0-kube-api-access-fv4s2\") pod \"redhat-marketplace-7cm84\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.564630 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.634600 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mzln7"] Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.641498 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.652449 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzln7"] Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.773406 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kvpd\" (UniqueName: \"kubernetes.io/projected/fe403490-5471-4d5f-a032-29f0325f8ac3-kube-api-access-6kvpd\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.773721 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-utilities\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.773748 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-catalog-content\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.783401 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cm84"] Oct 02 07:19:19 crc kubenswrapper[4829]: W1002 07:19:19.792317 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37c0917f_a686_4dc9_99d3_28ff284ca2e0.slice/crio-786eaed23ca90890286d08435499c33a93d64775a4934dc452e727ff0e5a7f43 WatchSource:0}: Error finding container 786eaed23ca90890286d08435499c33a93d64775a4934dc452e727ff0e5a7f43: Status 404 returned error can't find the container with id 786eaed23ca90890286d08435499c33a93d64775a4934dc452e727ff0e5a7f43 Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.874797 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-utilities\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.874849 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-catalog-content\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.874899 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kvpd\" (UniqueName: \"kubernetes.io/projected/fe403490-5471-4d5f-a032-29f0325f8ac3-kube-api-access-6kvpd\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.875420 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-catalog-content\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.875532 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-utilities\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.905981 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kvpd\" (UniqueName: \"kubernetes.io/projected/fe403490-5471-4d5f-a032-29f0325f8ac3-kube-api-access-6kvpd\") pod \"redhat-marketplace-mzln7\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:19 crc kubenswrapper[4829]: I1002 07:19:19.961656 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.135879 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzln7"] Oct 02 07:19:20 crc kubenswrapper[4829]: W1002 07:19:20.168960 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe403490_5471_4d5f_a032_29f0325f8ac3.slice/crio-f73b38900b5edbcf0fab3ce2709e297214d923f2f2ce52fa874e41c5d632d4e9 WatchSource:0}: Error finding container f73b38900b5edbcf0fab3ce2709e297214d923f2f2ce52fa874e41c5d632d4e9: Status 404 returned error can't find the container with id f73b38900b5edbcf0fab3ce2709e297214d923f2f2ce52fa874e41c5d632d4e9 Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.239985 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6hj5v"] Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.242082 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.244087 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.248875 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hj5v"] Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.381076 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5598\" (UniqueName: \"kubernetes.io/projected/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-kube-api-access-z5598\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.381127 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-utilities\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.381158 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-catalog-content\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.387129 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:20 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:20 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:20 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.387198 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.424081 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" event={"ID":"9c41ceaa-bc0a-4576-b309-0e7679130b71","Type":"ContainerStarted","Data":"b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.424130 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" event={"ID":"9c41ceaa-bc0a-4576-b309-0e7679130b71","Type":"ContainerStarted","Data":"2c85a608ab8876f259c8016732286b1ed45853ab3768844cb8290517ed832c61"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.424286 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.436152 4829 generic.go:334] "Generic (PLEG): container finished" podID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerID="709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8" exitCode=0 Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.436286 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cm84" event={"ID":"37c0917f-a686-4dc9-99d3-28ff284ca2e0","Type":"ContainerDied","Data":"709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.436314 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cm84" event={"ID":"37c0917f-a686-4dc9-99d3-28ff284ca2e0","Type":"ContainerStarted","Data":"786eaed23ca90890286d08435499c33a93d64775a4934dc452e727ff0e5a7f43"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.440921 4829 generic.go:334] "Generic (PLEG): container finished" podID="21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" containerID="30c219a07ea7a3e8f0dd4125c1dd8f2a58047f3194b0c451a53d8875cd1eee71" exitCode=0 Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.441001 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" event={"ID":"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a","Type":"ContainerDied","Data":"30c219a07ea7a3e8f0dd4125c1dd8f2a58047f3194b0c451a53d8875cd1eee71"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.460395 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" podStartSLOduration=131.460377192 podStartE2EDuration="2m11.460377192s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:20.449912864 +0000 UTC m=+151.789561269" watchObservedRunningTime="2025-10-02 07:19:20.460377192 +0000 UTC m=+151.800025597" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.466789 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerStarted","Data":"f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.466831 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerStarted","Data":"f73b38900b5edbcf0fab3ce2709e297214d923f2f2ce52fa874e41c5d632d4e9"} Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.482064 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5598\" (UniqueName: \"kubernetes.io/projected/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-kube-api-access-z5598\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.482107 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-utilities\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.482127 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-catalog-content\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.482539 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-catalog-content\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.482745 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-utilities\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.532368 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5598\" (UniqueName: \"kubernetes.io/projected/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-kube-api-access-z5598\") pod \"redhat-operators-6hj5v\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.565193 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.582149 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.582972 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.592057 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.592356 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.600766 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.638294 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l2qvh"] Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.657954 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l2qvh"] Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.658149 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.690961 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.691854 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24gkt\" (UniqueName: \"kubernetes.io/projected/3b1f096a-3ff8-407c-87ef-d76e370aee43-kube-api-access-24gkt\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.691918 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.691951 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-catalog-content\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.691967 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-utilities\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.793587 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-utilities\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.793650 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.793684 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24gkt\" (UniqueName: \"kubernetes.io/projected/3b1f096a-3ff8-407c-87ef-d76e370aee43-kube-api-access-24gkt\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.793739 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.793765 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-catalog-content\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.794577 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-utilities\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.794626 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.798095 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-catalog-content\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.811354 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24gkt\" (UniqueName: \"kubernetes.io/projected/3b1f096a-3ff8-407c-87ef-d76e370aee43-kube-api-access-24gkt\") pod \"redhat-operators-l2qvh\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.828725 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.843512 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6hj5v"] Oct 02 07:19:20 crc kubenswrapper[4829]: W1002 07:19:20.881868 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf210ccef_7e77_4e8c_97ca_d2e3abeeaa55.slice/crio-353c270ee08312529e310e96e4fe4e4d254fbd318c728db10f122cc03c77a8be WatchSource:0}: Error finding container 353c270ee08312529e310e96e4fe4e4d254fbd318c728db10f122cc03c77a8be: Status 404 returned error can't find the container with id 353c270ee08312529e310e96e4fe4e4d254fbd318c728db10f122cc03c77a8be Oct 02 07:19:20 crc kubenswrapper[4829]: I1002 07:19:20.941658 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.001662 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.036031 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.036082 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.037544 4829 patch_prober.go:28] interesting pod/console-f9d7485db-2tvsb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.037609 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2tvsb" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.228867 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.309138 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l2qvh"] Oct 02 07:19:21 crc kubenswrapper[4829]: W1002 07:19:21.345262 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1f096a_3ff8_407c_87ef_d76e370aee43.slice/crio-251e0b812cb01e4b335fced5359bfc8c37021b193e24264bcfc8408378287e56 WatchSource:0}: Error finding container 251e0b812cb01e4b335fced5359bfc8c37021b193e24264bcfc8408378287e56: Status 404 returned error can't find the container with id 251e0b812cb01e4b335fced5359bfc8c37021b193e24264bcfc8408378287e56 Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.385168 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.389456 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:21 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:21 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:21 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.389535 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.389809 4829 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7cxf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.389852 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-r7cxf" podUID="ca598dbb-d00b-4361-88c0-34c22ad2be6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.389943 4829 patch_prober.go:28] interesting pod/downloads-7954f5f757-r7cxf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" start-of-body= Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.390009 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-r7cxf" podUID="ca598dbb-d00b-4361-88c0-34c22ad2be6e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.24:8080/\": dial tcp 10.217.0.24:8080: connect: connection refused" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.489615 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd","Type":"ContainerStarted","Data":"15bd0cb4906d4ffc846472ef74b69df7c3a22eee3b52895643a35045e7600d07"} Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.498052 4829 generic.go:334] "Generic (PLEG): container finished" podID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerID="1be1992547c15e9708e6605b74f491927916ad5e1a129016ffbc81d2b9a3d873" exitCode=0 Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.498128 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hj5v" event={"ID":"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55","Type":"ContainerDied","Data":"1be1992547c15e9708e6605b74f491927916ad5e1a129016ffbc81d2b9a3d873"} Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.498157 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hj5v" event={"ID":"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55","Type":"ContainerStarted","Data":"353c270ee08312529e310e96e4fe4e4d254fbd318c728db10f122cc03c77a8be"} Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.507138 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2qvh" event={"ID":"3b1f096a-3ff8-407c-87ef-d76e370aee43","Type":"ContainerStarted","Data":"251e0b812cb01e4b335fced5359bfc8c37021b193e24264bcfc8408378287e56"} Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.511785 4829 generic.go:334] "Generic (PLEG): container finished" podID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerID="f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1" exitCode=0 Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.512043 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerDied","Data":"f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1"} Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.732994 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.812132 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9pqw\" (UniqueName: \"kubernetes.io/projected/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-kube-api-access-r9pqw\") pod \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.812216 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-config-volume\") pod \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.812254 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-secret-volume\") pod \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\" (UID: \"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a\") " Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.823557 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-config-volume" (OuterVolumeSpecName: "config-volume") pod "21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" (UID: "21f59a4c-e5a4-4b81-8c74-82d7cd1db54a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.828012 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" (UID: "21f59a4c-e5a4-4b81-8c74-82d7cd1db54a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.837528 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-kube-api-access-r9pqw" (OuterVolumeSpecName: "kube-api-access-r9pqw") pod "21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" (UID: "21f59a4c-e5a4-4b81-8c74-82d7cd1db54a"). InnerVolumeSpecName "kube-api-access-r9pqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.840516 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.913957 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9pqw\" (UniqueName: \"kubernetes.io/projected/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-kube-api-access-r9pqw\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.913993 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:21 crc kubenswrapper[4829]: I1002 07:19:21.914051 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.387658 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:22 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:22 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:22 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.387716 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.541110 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd","Type":"ContainerStarted","Data":"954147211a2f28da287bb86b9f7edb03aa2688e940cbcbf4ca4f39ae287b483a"} Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.557271 4829 generic.go:334] "Generic (PLEG): container finished" podID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerID="f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861" exitCode=0 Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.557341 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2qvh" event={"ID":"3b1f096a-3ff8-407c-87ef-d76e370aee43","Type":"ContainerDied","Data":"f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861"} Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.566620 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" event={"ID":"21f59a4c-e5a4-4b81-8c74-82d7cd1db54a","Type":"ContainerDied","Data":"4056463d6bd304ccbd92d9ea70201ef90ae709ecd63d7fa6ee5f6469510a64ba"} Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.566664 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs" Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.566677 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4056463d6bd304ccbd92d9ea70201ef90ae709ecd63d7fa6ee5f6469510a64ba" Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.620040 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:22 crc kubenswrapper[4829]: I1002 07:19:22.626293 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-l2x2j" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.125887 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 07:19:23 crc kubenswrapper[4829]: E1002 07:19:23.126131 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" containerName="collect-profiles" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.126143 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" containerName="collect-profiles" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.126233 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" containerName="collect-profiles" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.126578 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.128828 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.128898 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.135669 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a759cf3a-59bf-478f-a87e-27165a2644e6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.135745 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a759cf3a-59bf-478f-a87e-27165a2644e6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.143441 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.237023 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a759cf3a-59bf-478f-a87e-27165a2644e6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.237114 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a759cf3a-59bf-478f-a87e-27165a2644e6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.237172 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a759cf3a-59bf-478f-a87e-27165a2644e6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.265461 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a759cf3a-59bf-478f-a87e-27165a2644e6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.386517 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:23 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:23 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:23 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.386775 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.507992 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.598700 4829 generic.go:334] "Generic (PLEG): container finished" podID="ffc175c8-2ee3-4e99-9ea2-650a50d81cfd" containerID="954147211a2f28da287bb86b9f7edb03aa2688e940cbcbf4ca4f39ae287b483a" exitCode=0 Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.598796 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd","Type":"ContainerDied","Data":"954147211a2f28da287bb86b9f7edb03aa2688e940cbcbf4ca4f39ae287b483a"} Oct 02 07:19:23 crc kubenswrapper[4829]: I1002 07:19:23.655925 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-qbhhx" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.013798 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.063677 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kube-api-access\") pod \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.063793 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kubelet-dir\") pod \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\" (UID: \"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd\") " Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.064219 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ffc175c8-2ee3-4e99-9ea2-650a50d81cfd" (UID: "ffc175c8-2ee3-4e99-9ea2-650a50d81cfd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.065137 4829 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.081597 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ffc175c8-2ee3-4e99-9ea2-650a50d81cfd" (UID: "ffc175c8-2ee3-4e99-9ea2-650a50d81cfd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.084146 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.152549 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.167290 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ffc175c8-2ee3-4e99-9ea2-650a50d81cfd-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.387280 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:24 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:24 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:24 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.387631 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.617240 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a759cf3a-59bf-478f-a87e-27165a2644e6","Type":"ContainerStarted","Data":"5c1d779d61e68e100cf640b057ddfa605fda44ddc971639f2e73f5da704ca5e1"} Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.654732 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ffc175c8-2ee3-4e99-9ea2-650a50d81cfd","Type":"ContainerDied","Data":"15bd0cb4906d4ffc846472ef74b69df7c3a22eee3b52895643a35045e7600d07"} Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.654789 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15bd0cb4906d4ffc846472ef74b69df7c3a22eee3b52895643a35045e7600d07" Oct 02 07:19:24 crc kubenswrapper[4829]: I1002 07:19:24.654814 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 07:19:25 crc kubenswrapper[4829]: I1002 07:19:25.329566 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:19:25 crc kubenswrapper[4829]: I1002 07:19:25.329921 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:19:25 crc kubenswrapper[4829]: I1002 07:19:25.385803 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:25 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:25 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:25 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:25 crc kubenswrapper[4829]: I1002 07:19:25.385875 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:25 crc kubenswrapper[4829]: I1002 07:19:25.665366 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a759cf3a-59bf-478f-a87e-27165a2644e6","Type":"ContainerStarted","Data":"7c68448836b244171198d63e7cf41a86c88ae7798f3a22ed0771d85a50663664"} Oct 02 07:19:25 crc kubenswrapper[4829]: I1002 07:19:25.694608 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.694577538 podStartE2EDuration="2.694577538s" podCreationTimestamp="2025-10-02 07:19:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:25.680128251 +0000 UTC m=+157.019776646" watchObservedRunningTime="2025-10-02 07:19:25.694577538 +0000 UTC m=+157.034225943" Oct 02 07:19:26 crc kubenswrapper[4829]: I1002 07:19:26.385826 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:26 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:26 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:26 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:26 crc kubenswrapper[4829]: I1002 07:19:26.385882 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:26 crc kubenswrapper[4829]: I1002 07:19:26.674060 4829 generic.go:334] "Generic (PLEG): container finished" podID="a759cf3a-59bf-478f-a87e-27165a2644e6" containerID="7c68448836b244171198d63e7cf41a86c88ae7798f3a22ed0771d85a50663664" exitCode=0 Oct 02 07:19:26 crc kubenswrapper[4829]: I1002 07:19:26.674168 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a759cf3a-59bf-478f-a87e-27165a2644e6","Type":"ContainerDied","Data":"7c68448836b244171198d63e7cf41a86c88ae7798f3a22ed0771d85a50663664"} Oct 02 07:19:27 crc kubenswrapper[4829]: I1002 07:19:27.388191 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:27 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:27 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:27 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:27 crc kubenswrapper[4829]: I1002 07:19:27.388255 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:28 crc kubenswrapper[4829]: I1002 07:19:28.386087 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:28 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:28 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:28 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:28 crc kubenswrapper[4829]: I1002 07:19:28.386158 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:29 crc kubenswrapper[4829]: I1002 07:19:29.387702 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:29 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:29 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:29 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:29 crc kubenswrapper[4829]: I1002 07:19:29.387781 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:30 crc kubenswrapper[4829]: I1002 07:19:30.386020 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:30 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:30 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:30 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:30 crc kubenswrapper[4829]: I1002 07:19:30.386578 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:31 crc kubenswrapper[4829]: I1002 07:19:31.036563 4829 patch_prober.go:28] interesting pod/console-f9d7485db-2tvsb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 02 07:19:31 crc kubenswrapper[4829]: I1002 07:19:31.036619 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2tvsb" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 02 07:19:31 crc kubenswrapper[4829]: I1002 07:19:31.384817 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:31 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:31 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:31 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:31 crc kubenswrapper[4829]: I1002 07:19:31.384873 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:31 crc kubenswrapper[4829]: I1002 07:19:31.395033 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-r7cxf" Oct 02 07:19:32 crc kubenswrapper[4829]: I1002 07:19:32.387295 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:32 crc kubenswrapper[4829]: [-]has-synced failed: reason withheld Oct 02 07:19:32 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:32 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:32 crc kubenswrapper[4829]: I1002 07:19:32.387490 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.182866 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.202644 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a759cf3a-59bf-478f-a87e-27165a2644e6-kube-api-access\") pod \"a759cf3a-59bf-478f-a87e-27165a2644e6\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.204578 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a759cf3a-59bf-478f-a87e-27165a2644e6-kubelet-dir\") pod \"a759cf3a-59bf-478f-a87e-27165a2644e6\" (UID: \"a759cf3a-59bf-478f-a87e-27165a2644e6\") " Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.204706 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a759cf3a-59bf-478f-a87e-27165a2644e6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a759cf3a-59bf-478f-a87e-27165a2644e6" (UID: "a759cf3a-59bf-478f-a87e-27165a2644e6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.205388 4829 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a759cf3a-59bf-478f-a87e-27165a2644e6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.256353 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a759cf3a-59bf-478f-a87e-27165a2644e6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a759cf3a-59bf-478f-a87e-27165a2644e6" (UID: "a759cf3a-59bf-478f-a87e-27165a2644e6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.307271 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a759cf3a-59bf-478f-a87e-27165a2644e6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.386384 4829 patch_prober.go:28] interesting pod/router-default-5444994796-cmbx8 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 07:19:33 crc kubenswrapper[4829]: [+]has-synced ok Oct 02 07:19:33 crc kubenswrapper[4829]: [+]process-running ok Oct 02 07:19:33 crc kubenswrapper[4829]: healthz check failed Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.386493 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cmbx8" podUID="12133a89-e03b-4a3f-b665-c5d790315e98" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.612121 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.624970 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/675c4639-4b87-404e-8258-0e5bae51d933-metrics-certs\") pod \"network-metrics-daemon-mxmxh\" (UID: \"675c4639-4b87-404e-8258-0e5bae51d933\") " pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.736645 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a759cf3a-59bf-478f-a87e-27165a2644e6","Type":"ContainerDied","Data":"5c1d779d61e68e100cf640b057ddfa605fda44ddc971639f2e73f5da704ca5e1"} Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.736702 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c1d779d61e68e100cf640b057ddfa605fda44ddc971639f2e73f5da704ca5e1" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.736742 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 07:19:33 crc kubenswrapper[4829]: I1002 07:19:33.778826 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mxmxh" Oct 02 07:19:34 crc kubenswrapper[4829]: I1002 07:19:34.386494 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:34 crc kubenswrapper[4829]: I1002 07:19:34.392002 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cmbx8" Oct 02 07:19:39 crc kubenswrapper[4829]: I1002 07:19:39.104874 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:19:41 crc kubenswrapper[4829]: I1002 07:19:41.043480 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:41 crc kubenswrapper[4829]: I1002 07:19:41.051256 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:19:51 crc kubenswrapper[4829]: I1002 07:19:51.805802 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jzdnq" Oct 02 07:19:53 crc kubenswrapper[4829]: E1002 07:19:53.630648 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 07:19:53 crc kubenswrapper[4829]: E1002 07:19:53.631056 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gb9kg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-7rgt2_openshift-marketplace(0357590d-1dad-4ef3-b422-8b25a71953ab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:53 crc kubenswrapper[4829]: E1002 07:19:53.632358 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-7rgt2" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" Oct 02 07:19:53 crc kubenswrapper[4829]: E1002 07:19:53.637531 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 07:19:53 crc kubenswrapper[4829]: E1002 07:19:53.637692 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-77k7x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rltqp_openshift-marketplace(330c4c1e-cb32-42fd-a3a8-8256339e5e17): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:53 crc kubenswrapper[4829]: E1002 07:19:53.639064 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rltqp" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.822022 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rltqp" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.822095 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-7rgt2" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.889619 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.889776 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnq4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-xl9qq_openshift-marketplace(3cd1d972-be99-4ae8-b965-dd55a308e3a9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.892701 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-xl9qq" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.912957 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.913108 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dhd7n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-4lkc7_openshift-marketplace(8d3333e7-33d0-4f71-b051-6980f83a77a4): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:54 crc kubenswrapper[4829]: E1002 07:19:54.914409 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-4lkc7" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" Oct 02 07:19:55 crc kubenswrapper[4829]: I1002 07:19:55.328909 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:19:55 crc kubenswrapper[4829]: I1002 07:19:55.328956 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:19:57 crc kubenswrapper[4829]: I1002 07:19:57.498533 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 07:19:57 crc kubenswrapper[4829]: E1002 07:19:57.617355 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-xl9qq" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" Oct 02 07:19:57 crc kubenswrapper[4829]: E1002 07:19:57.617628 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-4lkc7" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" Oct 02 07:19:57 crc kubenswrapper[4829]: E1002 07:19:57.677250 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 07:19:57 crc kubenswrapper[4829]: E1002 07:19:57.677388 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-24gkt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-l2qvh_openshift-marketplace(3b1f096a-3ff8-407c-87ef-d76e370aee43): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:57 crc kubenswrapper[4829]: E1002 07:19:57.678703 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-l2qvh" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" Oct 02 07:19:57 crc kubenswrapper[4829]: I1002 07:19:57.976751 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mxmxh"] Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.400301 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-l2qvh" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.458177 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.458631 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fv4s2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7cm84_openshift-marketplace(37c0917f-a686-4dc9-99d3-28ff284ca2e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.460428 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7cm84" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.479660 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.479868 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5598,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-6hj5v_openshift-marketplace(f210ccef-7e77-4e8c-97ca-d2e3abeeaa55): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.481332 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-6hj5v" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.499404 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.499565 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kvpd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mzln7_openshift-marketplace(fe403490-5471-4d5f-a032-29f0325f8ac3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.501463 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mzln7" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" Oct 02 07:19:58 crc kubenswrapper[4829]: I1002 07:19:58.883317 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" event={"ID":"675c4639-4b87-404e-8258-0e5bae51d933","Type":"ContainerStarted","Data":"b2d99862ca6e420f951db9705e2a04c4ca1c76c8ee3ee94bb4b59d304ad14f76"} Oct 02 07:19:58 crc kubenswrapper[4829]: I1002 07:19:58.883370 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" event={"ID":"675c4639-4b87-404e-8258-0e5bae51d933","Type":"ContainerStarted","Data":"d2d30e3bf9fd4297646ae92252be8cffc55a4e072767b98d735e5e03fb0adc45"} Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.884680 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-6hj5v" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.884905 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7cm84" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" Oct 02 07:19:58 crc kubenswrapper[4829]: E1002 07:19:58.889114 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mzln7" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" Oct 02 07:19:59 crc kubenswrapper[4829]: I1002 07:19:59.891207 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mxmxh" event={"ID":"675c4639-4b87-404e-8258-0e5bae51d933","Type":"ContainerStarted","Data":"8fcbb0a123d318aff3e89ef47fe7b89d302cfada093ddd9f8d76800bbd556c87"} Oct 02 07:19:59 crc kubenswrapper[4829]: I1002 07:19:59.921574 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-mxmxh" podStartSLOduration=170.92154173 podStartE2EDuration="2m50.92154173s" podCreationTimestamp="2025-10-02 07:17:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:19:59.913717411 +0000 UTC m=+191.253365846" watchObservedRunningTime="2025-10-02 07:19:59.92154173 +0000 UTC m=+191.261190165" Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.979098 4829 generic.go:334] "Generic (PLEG): container finished" podID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerID="7680fe23e4b068b7198564f6aff3fd576d57eb20a8e229ff1c6134e8f4f86cf4" exitCode=0 Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.979172 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hj5v" event={"ID":"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55","Type":"ContainerDied","Data":"7680fe23e4b068b7198564f6aff3fd576d57eb20a8e229ff1c6134e8f4f86cf4"} Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.984042 4829 generic.go:334] "Generic (PLEG): container finished" podID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerID="92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341" exitCode=0 Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.984102 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rgt2" event={"ID":"0357590d-1dad-4ef3-b422-8b25a71953ab","Type":"ContainerDied","Data":"92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341"} Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.988341 4829 generic.go:334] "Generic (PLEG): container finished" podID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerID="fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039" exitCode=0 Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.988408 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rltqp" event={"ID":"330c4c1e-cb32-42fd-a3a8-8256339e5e17","Type":"ContainerDied","Data":"fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039"} Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.991974 4829 generic.go:334] "Generic (PLEG): container finished" podID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerID="328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486" exitCode=0 Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.992033 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lkc7" event={"ID":"8d3333e7-33d0-4f71-b051-6980f83a77a4","Type":"ContainerDied","Data":"328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486"} Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.993945 4829 generic.go:334] "Generic (PLEG): container finished" podID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerID="3af68e7aedcff8a48f0ba5232eba7ed9aa453f030a85578b31eb062b36997827" exitCode=0 Oct 02 07:20:13 crc kubenswrapper[4829]: I1002 07:20:13.993966 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl9qq" event={"ID":"3cd1d972-be99-4ae8-b965-dd55a308e3a9","Type":"ContainerDied","Data":"3af68e7aedcff8a48f0ba5232eba7ed9aa453f030a85578b31eb062b36997827"} Oct 02 07:20:15 crc kubenswrapper[4829]: I1002 07:20:15.004345 4829 generic.go:334] "Generic (PLEG): container finished" podID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerID="4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e" exitCode=0 Oct 02 07:20:15 crc kubenswrapper[4829]: I1002 07:20:15.004410 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2qvh" event={"ID":"3b1f096a-3ff8-407c-87ef-d76e370aee43","Type":"ContainerDied","Data":"4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e"} Oct 02 07:20:15 crc kubenswrapper[4829]: I1002 07:20:15.009737 4829 generic.go:334] "Generic (PLEG): container finished" podID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerID="a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b" exitCode=0 Oct 02 07:20:15 crc kubenswrapper[4829]: I1002 07:20:15.009790 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cm84" event={"ID":"37c0917f-a686-4dc9-99d3-28ff284ca2e0","Type":"ContainerDied","Data":"a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.017798 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2qvh" event={"ID":"3b1f096a-3ff8-407c-87ef-d76e370aee43","Type":"ContainerStarted","Data":"29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.019280 4829 generic.go:334] "Generic (PLEG): container finished" podID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerID="773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36" exitCode=0 Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.019349 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerDied","Data":"773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.022326 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl9qq" event={"ID":"3cd1d972-be99-4ae8-b965-dd55a308e3a9","Type":"ContainerStarted","Data":"ae6cdc959d27b6a3edaac0f0dff7d2a325452ba2786bf2ff4c052ba7874f2882"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.024833 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hj5v" event={"ID":"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55","Type":"ContainerStarted","Data":"c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.027312 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rgt2" event={"ID":"0357590d-1dad-4ef3-b422-8b25a71953ab","Type":"ContainerStarted","Data":"a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.030433 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rltqp" event={"ID":"330c4c1e-cb32-42fd-a3a8-8256339e5e17","Type":"ContainerStarted","Data":"ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.032982 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cm84" event={"ID":"37c0917f-a686-4dc9-99d3-28ff284ca2e0","Type":"ContainerStarted","Data":"7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.035736 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lkc7" event={"ID":"8d3333e7-33d0-4f71-b051-6980f83a77a4","Type":"ContainerStarted","Data":"c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646"} Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.093596 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6hj5v" podStartSLOduration=2.624013024 podStartE2EDuration="56.093582511s" podCreationTimestamp="2025-10-02 07:19:20 +0000 UTC" firstStartedPulling="2025-10-02 07:19:21.507093064 +0000 UTC m=+152.846741459" lastFinishedPulling="2025-10-02 07:20:14.976662541 +0000 UTC m=+206.316310946" observedRunningTime="2025-10-02 07:20:16.09059645 +0000 UTC m=+207.430244855" watchObservedRunningTime="2025-10-02 07:20:16.093582511 +0000 UTC m=+207.433230916" Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.093939 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l2qvh" podStartSLOduration=3.080112843 podStartE2EDuration="56.093933298s" podCreationTimestamp="2025-10-02 07:19:20 +0000 UTC" firstStartedPulling="2025-10-02 07:19:22.569775565 +0000 UTC m=+153.909423970" lastFinishedPulling="2025-10-02 07:20:15.58359602 +0000 UTC m=+206.923244425" observedRunningTime="2025-10-02 07:20:16.073517795 +0000 UTC m=+207.413166200" watchObservedRunningTime="2025-10-02 07:20:16.093933298 +0000 UTC m=+207.433581703" Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.108683 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xl9qq" podStartSLOduration=3.6029356999999997 podStartE2EDuration="59.108669499s" podCreationTimestamp="2025-10-02 07:19:17 +0000 UTC" firstStartedPulling="2025-10-02 07:19:19.378531576 +0000 UTC m=+150.718179971" lastFinishedPulling="2025-10-02 07:20:14.884265365 +0000 UTC m=+206.223913770" observedRunningTime="2025-10-02 07:20:16.105302159 +0000 UTC m=+207.444950564" watchObservedRunningTime="2025-10-02 07:20:16.108669499 +0000 UTC m=+207.448317904" Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.125859 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7cm84" podStartSLOduration=2.122408992 podStartE2EDuration="57.125844448s" podCreationTimestamp="2025-10-02 07:19:19 +0000 UTC" firstStartedPulling="2025-10-02 07:19:20.437668323 +0000 UTC m=+151.777316728" lastFinishedPulling="2025-10-02 07:20:15.441103769 +0000 UTC m=+206.780752184" observedRunningTime="2025-10-02 07:20:16.122393185 +0000 UTC m=+207.462041580" watchObservedRunningTime="2025-10-02 07:20:16.125844448 +0000 UTC m=+207.465492853" Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.142972 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4lkc7" podStartSLOduration=3.485879869 podStartE2EDuration="59.142956605s" podCreationTimestamp="2025-10-02 07:19:17 +0000 UTC" firstStartedPulling="2025-10-02 07:19:19.357826936 +0000 UTC m=+150.697475341" lastFinishedPulling="2025-10-02 07:20:15.014903672 +0000 UTC m=+206.354552077" observedRunningTime="2025-10-02 07:20:16.140447374 +0000 UTC m=+207.480095789" watchObservedRunningTime="2025-10-02 07:20:16.142956605 +0000 UTC m=+207.482605010" Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.161703 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7rgt2" podStartSLOduration=3.643335695 podStartE2EDuration="59.161687933s" podCreationTimestamp="2025-10-02 07:19:17 +0000 UTC" firstStartedPulling="2025-10-02 07:19:19.391367144 +0000 UTC m=+150.731015549" lastFinishedPulling="2025-10-02 07:20:14.909719382 +0000 UTC m=+206.249367787" observedRunningTime="2025-10-02 07:20:16.158533734 +0000 UTC m=+207.498182139" watchObservedRunningTime="2025-10-02 07:20:16.161687933 +0000 UTC m=+207.501336338" Oct 02 07:20:16 crc kubenswrapper[4829]: I1002 07:20:16.179089 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rltqp" podStartSLOduration=3.522726967 podStartE2EDuration="59.179069242s" podCreationTimestamp="2025-10-02 07:19:17 +0000 UTC" firstStartedPulling="2025-10-02 07:19:19.432314399 +0000 UTC m=+150.771962804" lastFinishedPulling="2025-10-02 07:20:15.088656674 +0000 UTC m=+206.428305079" observedRunningTime="2025-10-02 07:20:16.177823007 +0000 UTC m=+207.517471402" watchObservedRunningTime="2025-10-02 07:20:16.179069242 +0000 UTC m=+207.518717647" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.042023 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerStarted","Data":"8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1"} Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.065169 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mzln7" podStartSLOduration=1.831730958 podStartE2EDuration="58.065151925s" podCreationTimestamp="2025-10-02 07:19:19 +0000 UTC" firstStartedPulling="2025-10-02 07:19:20.474408056 +0000 UTC m=+151.814056461" lastFinishedPulling="2025-10-02 07:20:16.707829023 +0000 UTC m=+208.047477428" observedRunningTime="2025-10-02 07:20:17.062290548 +0000 UTC m=+208.401938953" watchObservedRunningTime="2025-10-02 07:20:17.065151925 +0000 UTC m=+208.404800330" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.385713 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.385772 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.594661 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.594920 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.843943 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:20:17 crc kubenswrapper[4829]: I1002 07:20:17.843987 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.054697 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.054742 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.095078 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.536417 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-7rgt2" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="registry-server" probeResult="failure" output=< Oct 02 07:20:18 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:20:18 crc kubenswrapper[4829]: > Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.633816 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-xl9qq" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="registry-server" probeResult="failure" output=< Oct 02 07:20:18 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:20:18 crc kubenswrapper[4829]: > Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.756721 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6w49n"] Oct 02 07:20:18 crc kubenswrapper[4829]: I1002 07:20:18.888751 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rltqp" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="registry-server" probeResult="failure" output=< Oct 02 07:20:18 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:20:18 crc kubenswrapper[4829]: > Oct 02 07:20:19 crc kubenswrapper[4829]: I1002 07:20:19.565078 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:20:19 crc kubenswrapper[4829]: I1002 07:20:19.565120 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:20:19 crc kubenswrapper[4829]: I1002 07:20:19.618656 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:20:19 crc kubenswrapper[4829]: I1002 07:20:19.963073 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:20:19 crc kubenswrapper[4829]: I1002 07:20:19.963419 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:20:20 crc kubenswrapper[4829]: I1002 07:20:20.001151 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:20:20 crc kubenswrapper[4829]: I1002 07:20:20.097581 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:20:20 crc kubenswrapper[4829]: I1002 07:20:20.566010 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:20:20 crc kubenswrapper[4829]: I1002 07:20:20.566062 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:20:20 crc kubenswrapper[4829]: I1002 07:20:20.607978 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:20:21 crc kubenswrapper[4829]: I1002 07:20:21.002190 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:20:21 crc kubenswrapper[4829]: I1002 07:20:21.002250 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:20:21 crc kubenswrapper[4829]: I1002 07:20:21.064000 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:20:21 crc kubenswrapper[4829]: I1002 07:20:21.103664 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:20:21 crc kubenswrapper[4829]: I1002 07:20:21.118294 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:20:21 crc kubenswrapper[4829]: I1002 07:20:21.133754 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.305568 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l2qvh"] Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.306266 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l2qvh" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="registry-server" containerID="cri-o://29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3" gracePeriod=2 Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.716381 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.866766 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-utilities\") pod \"3b1f096a-3ff8-407c-87ef-d76e370aee43\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.867122 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-catalog-content\") pod \"3b1f096a-3ff8-407c-87ef-d76e370aee43\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.867208 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24gkt\" (UniqueName: \"kubernetes.io/projected/3b1f096a-3ff8-407c-87ef-d76e370aee43-kube-api-access-24gkt\") pod \"3b1f096a-3ff8-407c-87ef-d76e370aee43\" (UID: \"3b1f096a-3ff8-407c-87ef-d76e370aee43\") " Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.867516 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-utilities" (OuterVolumeSpecName: "utilities") pod "3b1f096a-3ff8-407c-87ef-d76e370aee43" (UID: "3b1f096a-3ff8-407c-87ef-d76e370aee43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.867708 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.873093 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1f096a-3ff8-407c-87ef-d76e370aee43-kube-api-access-24gkt" (OuterVolumeSpecName: "kube-api-access-24gkt") pod "3b1f096a-3ff8-407c-87ef-d76e370aee43" (UID: "3b1f096a-3ff8-407c-87ef-d76e370aee43"). InnerVolumeSpecName "kube-api-access-24gkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.953837 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b1f096a-3ff8-407c-87ef-d76e370aee43" (UID: "3b1f096a-3ff8-407c-87ef-d76e370aee43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.970048 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1f096a-3ff8-407c-87ef-d76e370aee43-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:23 crc kubenswrapper[4829]: I1002 07:20:23.970088 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24gkt\" (UniqueName: \"kubernetes.io/projected/3b1f096a-3ff8-407c-87ef-d76e370aee43-kube-api-access-24gkt\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.095045 4829 generic.go:334] "Generic (PLEG): container finished" podID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerID="29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3" exitCode=0 Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.095125 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2qvh" event={"ID":"3b1f096a-3ff8-407c-87ef-d76e370aee43","Type":"ContainerDied","Data":"29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3"} Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.095153 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l2qvh" event={"ID":"3b1f096a-3ff8-407c-87ef-d76e370aee43","Type":"ContainerDied","Data":"251e0b812cb01e4b335fced5359bfc8c37021b193e24264bcfc8408378287e56"} Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.095185 4829 scope.go:117] "RemoveContainer" containerID="29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.095368 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l2qvh" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.116073 4829 scope.go:117] "RemoveContainer" containerID="4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.129680 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l2qvh"] Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.132106 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l2qvh"] Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.156592 4829 scope.go:117] "RemoveContainer" containerID="f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.170950 4829 scope.go:117] "RemoveContainer" containerID="29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3" Oct 02 07:20:24 crc kubenswrapper[4829]: E1002 07:20:24.171333 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3\": container with ID starting with 29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3 not found: ID does not exist" containerID="29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.171364 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3"} err="failed to get container status \"29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3\": rpc error: code = NotFound desc = could not find container \"29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3\": container with ID starting with 29a20cd884f870b5fe00d0586ce393f476792f8cd7969a71d1791ff951fea8c3 not found: ID does not exist" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.171411 4829 scope.go:117] "RemoveContainer" containerID="4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e" Oct 02 07:20:24 crc kubenswrapper[4829]: E1002 07:20:24.171628 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e\": container with ID starting with 4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e not found: ID does not exist" containerID="4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.171662 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e"} err="failed to get container status \"4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e\": rpc error: code = NotFound desc = could not find container \"4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e\": container with ID starting with 4073c0e71c6fc1756f4a70b90e72b8df7c75eeb31530ac52b7ad40004044af4e not found: ID does not exist" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.171683 4829 scope.go:117] "RemoveContainer" containerID="f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861" Oct 02 07:20:24 crc kubenswrapper[4829]: E1002 07:20:24.171846 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861\": container with ID starting with f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861 not found: ID does not exist" containerID="f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.171871 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861"} err="failed to get container status \"f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861\": rpc error: code = NotFound desc = could not find container \"f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861\": container with ID starting with f46dc3642eb7cf235ac3f046d6879bfa602390f8b27b182ee7ae49a7055cc861 not found: ID does not exist" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.302452 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzln7"] Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.302807 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mzln7" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="registry-server" containerID="cri-o://8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1" gracePeriod=2 Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.619754 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.777467 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kvpd\" (UniqueName: \"kubernetes.io/projected/fe403490-5471-4d5f-a032-29f0325f8ac3-kube-api-access-6kvpd\") pod \"fe403490-5471-4d5f-a032-29f0325f8ac3\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.777537 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-catalog-content\") pod \"fe403490-5471-4d5f-a032-29f0325f8ac3\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.777587 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-utilities\") pod \"fe403490-5471-4d5f-a032-29f0325f8ac3\" (UID: \"fe403490-5471-4d5f-a032-29f0325f8ac3\") " Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.779119 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-utilities" (OuterVolumeSpecName: "utilities") pod "fe403490-5471-4d5f-a032-29f0325f8ac3" (UID: "fe403490-5471-4d5f-a032-29f0325f8ac3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.785172 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe403490-5471-4d5f-a032-29f0325f8ac3-kube-api-access-6kvpd" (OuterVolumeSpecName: "kube-api-access-6kvpd") pod "fe403490-5471-4d5f-a032-29f0325f8ac3" (UID: "fe403490-5471-4d5f-a032-29f0325f8ac3"). InnerVolumeSpecName "kube-api-access-6kvpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.789279 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe403490-5471-4d5f-a032-29f0325f8ac3" (UID: "fe403490-5471-4d5f-a032-29f0325f8ac3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.879682 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kvpd\" (UniqueName: \"kubernetes.io/projected/fe403490-5471-4d5f-a032-29f0325f8ac3-kube-api-access-6kvpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.879715 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:24 crc kubenswrapper[4829]: I1002 07:20:24.879724 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe403490-5471-4d5f-a032-29f0325f8ac3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.105220 4829 generic.go:334] "Generic (PLEG): container finished" podID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerID="8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1" exitCode=0 Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.105318 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerDied","Data":"8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1"} Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.105325 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mzln7" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.105346 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mzln7" event={"ID":"fe403490-5471-4d5f-a032-29f0325f8ac3","Type":"ContainerDied","Data":"f73b38900b5edbcf0fab3ce2709e297214d923f2f2ce52fa874e41c5d632d4e9"} Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.105363 4829 scope.go:117] "RemoveContainer" containerID="8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.129977 4829 scope.go:117] "RemoveContainer" containerID="773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.146974 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzln7"] Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.150138 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mzln7"] Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.168537 4829 scope.go:117] "RemoveContainer" containerID="f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.186649 4829 scope.go:117] "RemoveContainer" containerID="8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1" Oct 02 07:20:25 crc kubenswrapper[4829]: E1002 07:20:25.186957 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1\": container with ID starting with 8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1 not found: ID does not exist" containerID="8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.186989 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1"} err="failed to get container status \"8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1\": rpc error: code = NotFound desc = could not find container \"8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1\": container with ID starting with 8727c372e4a17affae6b6ad0f7d526fb644c3b8e6d856919eb1f517b1bdd85a1 not found: ID does not exist" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.187009 4829 scope.go:117] "RemoveContainer" containerID="773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36" Oct 02 07:20:25 crc kubenswrapper[4829]: E1002 07:20:25.187320 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36\": container with ID starting with 773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36 not found: ID does not exist" containerID="773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.187463 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36"} err="failed to get container status \"773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36\": rpc error: code = NotFound desc = could not find container \"773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36\": container with ID starting with 773742df098e774e4a4b4a1ddd523f1936b95fc58ffacca2c642381288c85f36 not found: ID does not exist" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.187577 4829 scope.go:117] "RemoveContainer" containerID="f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1" Oct 02 07:20:25 crc kubenswrapper[4829]: E1002 07:20:25.188040 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1\": container with ID starting with f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1 not found: ID does not exist" containerID="f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.188064 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1"} err="failed to get container status \"f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1\": rpc error: code = NotFound desc = could not find container \"f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1\": container with ID starting with f73d010991d15a11338bd10cf36372f3f44586a06e8b92bb7c148789ea254af1 not found: ID does not exist" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.328992 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.329047 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.329097 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.329789 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.329861 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146" gracePeriod=600 Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.476203 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" path="/var/lib/kubelet/pods/3b1f096a-3ff8-407c-87ef-d76e370aee43/volumes" Oct 02 07:20:25 crc kubenswrapper[4829]: I1002 07:20:25.477055 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" path="/var/lib/kubelet/pods/fe403490-5471-4d5f-a032-29f0325f8ac3/volumes" Oct 02 07:20:26 crc kubenswrapper[4829]: I1002 07:20:26.115547 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146" exitCode=0 Oct 02 07:20:26 crc kubenswrapper[4829]: I1002 07:20:26.115681 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146"} Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.124142 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"eaf4b9893a38009f8a714f08d03494fe5d6ecc601b0a3a024d68f3487caee9a2"} Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.468840 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.518896 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.643528 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.695973 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.891650 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:20:27 crc kubenswrapper[4829]: I1002 07:20:27.949708 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:20:28 crc kubenswrapper[4829]: I1002 07:20:28.099809 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:20:29 crc kubenswrapper[4829]: I1002 07:20:29.701782 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rltqp"] Oct 02 07:20:29 crc kubenswrapper[4829]: I1002 07:20:29.702326 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rltqp" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="registry-server" containerID="cri-o://ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88" gracePeriod=2 Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.056560 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.138320 4829 generic.go:334] "Generic (PLEG): container finished" podID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerID="ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88" exitCode=0 Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.138356 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rltqp" event={"ID":"330c4c1e-cb32-42fd-a3a8-8256339e5e17","Type":"ContainerDied","Data":"ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88"} Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.138394 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rltqp" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.138409 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rltqp" event={"ID":"330c4c1e-cb32-42fd-a3a8-8256339e5e17","Type":"ContainerDied","Data":"a2285a7615f4768f6fb3ba1121494581344af3fa8ebccaf5aad0581967f4cb6c"} Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.138431 4829 scope.go:117] "RemoveContainer" containerID="ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.154687 4829 scope.go:117] "RemoveContainer" containerID="fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.164420 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-utilities\") pod \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.164485 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77k7x\" (UniqueName: \"kubernetes.io/projected/330c4c1e-cb32-42fd-a3a8-8256339e5e17-kube-api-access-77k7x\") pod \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.164520 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-catalog-content\") pod \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\" (UID: \"330c4c1e-cb32-42fd-a3a8-8256339e5e17\") " Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.165114 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-utilities" (OuterVolumeSpecName: "utilities") pod "330c4c1e-cb32-42fd-a3a8-8256339e5e17" (UID: "330c4c1e-cb32-42fd-a3a8-8256339e5e17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.169811 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330c4c1e-cb32-42fd-a3a8-8256339e5e17-kube-api-access-77k7x" (OuterVolumeSpecName: "kube-api-access-77k7x") pod "330c4c1e-cb32-42fd-a3a8-8256339e5e17" (UID: "330c4c1e-cb32-42fd-a3a8-8256339e5e17"). InnerVolumeSpecName "kube-api-access-77k7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.171307 4829 scope.go:117] "RemoveContainer" containerID="0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.207258 4829 scope.go:117] "RemoveContainer" containerID="ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88" Oct 02 07:20:30 crc kubenswrapper[4829]: E1002 07:20:30.211618 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88\": container with ID starting with ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88 not found: ID does not exist" containerID="ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.211664 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88"} err="failed to get container status \"ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88\": rpc error: code = NotFound desc = could not find container \"ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88\": container with ID starting with ec4c5c975625d39f7f95dede9eaf7805fa94f24e6697e11e4bd9cfddd62ddc88 not found: ID does not exist" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.211692 4829 scope.go:117] "RemoveContainer" containerID="fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039" Oct 02 07:20:30 crc kubenswrapper[4829]: E1002 07:20:30.213380 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039\": container with ID starting with fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039 not found: ID does not exist" containerID="fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.213408 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039"} err="failed to get container status \"fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039\": rpc error: code = NotFound desc = could not find container \"fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039\": container with ID starting with fd1e452e10cddd9e5901d1c864edff81095708326e0db089f226b54280f17039 not found: ID does not exist" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.213426 4829 scope.go:117] "RemoveContainer" containerID="0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82" Oct 02 07:20:30 crc kubenswrapper[4829]: E1002 07:20:30.214351 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82\": container with ID starting with 0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82 not found: ID does not exist" containerID="0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.214375 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82"} err="failed to get container status \"0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82\": rpc error: code = NotFound desc = could not find container \"0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82\": container with ID starting with 0e0a8f797fab4c64282be84acf46269d3c5f5c9e8b34d8ad6f83ff455b3a9c82 not found: ID does not exist" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.229260 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "330c4c1e-cb32-42fd-a3a8-8256339e5e17" (UID: "330c4c1e-cb32-42fd-a3a8-8256339e5e17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.265539 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77k7x\" (UniqueName: \"kubernetes.io/projected/330c4c1e-cb32-42fd-a3a8-8256339e5e17-kube-api-access-77k7x\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.265581 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.265596 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/330c4c1e-cb32-42fd-a3a8-8256339e5e17-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.487562 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rltqp"] Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.490342 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rltqp"] Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.700488 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4lkc7"] Oct 02 07:20:30 crc kubenswrapper[4829]: I1002 07:20:30.700972 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4lkc7" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="registry-server" containerID="cri-o://c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646" gracePeriod=2 Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.055770 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.086387 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-utilities\") pod \"8d3333e7-33d0-4f71-b051-6980f83a77a4\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.086450 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhd7n\" (UniqueName: \"kubernetes.io/projected/8d3333e7-33d0-4f71-b051-6980f83a77a4-kube-api-access-dhd7n\") pod \"8d3333e7-33d0-4f71-b051-6980f83a77a4\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.086554 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-catalog-content\") pod \"8d3333e7-33d0-4f71-b051-6980f83a77a4\" (UID: \"8d3333e7-33d0-4f71-b051-6980f83a77a4\") " Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.087047 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-utilities" (OuterVolumeSpecName: "utilities") pod "8d3333e7-33d0-4f71-b051-6980f83a77a4" (UID: "8d3333e7-33d0-4f71-b051-6980f83a77a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.091252 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d3333e7-33d0-4f71-b051-6980f83a77a4-kube-api-access-dhd7n" (OuterVolumeSpecName: "kube-api-access-dhd7n") pod "8d3333e7-33d0-4f71-b051-6980f83a77a4" (UID: "8d3333e7-33d0-4f71-b051-6980f83a77a4"). InnerVolumeSpecName "kube-api-access-dhd7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.135059 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d3333e7-33d0-4f71-b051-6980f83a77a4" (UID: "8d3333e7-33d0-4f71-b051-6980f83a77a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.145038 4829 generic.go:334] "Generic (PLEG): container finished" podID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerID="c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646" exitCode=0 Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.145077 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lkc7" event={"ID":"8d3333e7-33d0-4f71-b051-6980f83a77a4","Type":"ContainerDied","Data":"c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646"} Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.145110 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4lkc7" event={"ID":"8d3333e7-33d0-4f71-b051-6980f83a77a4","Type":"ContainerDied","Data":"995ebc9b1c60ae1b74c8ed0aff80dad26609f95ca5453903906db8d16ea535e1"} Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.145126 4829 scope.go:117] "RemoveContainer" containerID="c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.145122 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4lkc7" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.159911 4829 scope.go:117] "RemoveContainer" containerID="328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.171453 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4lkc7"] Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.177418 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4lkc7"] Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.188032 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.188064 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhd7n\" (UniqueName: \"kubernetes.io/projected/8d3333e7-33d0-4f71-b051-6980f83a77a4-kube-api-access-dhd7n\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.188076 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d3333e7-33d0-4f71-b051-6980f83a77a4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.199583 4829 scope.go:117] "RemoveContainer" containerID="b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.213218 4829 scope.go:117] "RemoveContainer" containerID="c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646" Oct 02 07:20:31 crc kubenswrapper[4829]: E1002 07:20:31.213798 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646\": container with ID starting with c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646 not found: ID does not exist" containerID="c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.213827 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646"} err="failed to get container status \"c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646\": rpc error: code = NotFound desc = could not find container \"c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646\": container with ID starting with c891bb2aba3e819eecfd988de6e28004e04968b18a688ea3265ea88f3e5ee646 not found: ID does not exist" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.213851 4829 scope.go:117] "RemoveContainer" containerID="328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486" Oct 02 07:20:31 crc kubenswrapper[4829]: E1002 07:20:31.214178 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486\": container with ID starting with 328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486 not found: ID does not exist" containerID="328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.214259 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486"} err="failed to get container status \"328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486\": rpc error: code = NotFound desc = could not find container \"328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486\": container with ID starting with 328ddc6332c3bfbcc56bad8205435c439a1eff71928e1e3efc6a03d2bdd00486 not found: ID does not exist" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.214312 4829 scope.go:117] "RemoveContainer" containerID="b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3" Oct 02 07:20:31 crc kubenswrapper[4829]: E1002 07:20:31.214710 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3\": container with ID starting with b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3 not found: ID does not exist" containerID="b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.214771 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3"} err="failed to get container status \"b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3\": rpc error: code = NotFound desc = could not find container \"b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3\": container with ID starting with b61be61b51695ecf3fb0a370a3eff93c7f8d7fc26d6f8cfbf782791aabe9a6e3 not found: ID does not exist" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.466122 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" path="/var/lib/kubelet/pods/330c4c1e-cb32-42fd-a3a8-8256339e5e17/volumes" Oct 02 07:20:31 crc kubenswrapper[4829]: I1002 07:20:31.466755 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" path="/var/lib/kubelet/pods/8d3333e7-33d0-4f71-b051-6980f83a77a4/volumes" Oct 02 07:20:43 crc kubenswrapper[4829]: I1002 07:20:43.787514 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerName="oauth-openshift" containerID="cri-o://ac66a22531034d0affa1529759b983ac2f8598f419b9bc97a55b6b616367e33c" gracePeriod=15 Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.225316 4829 generic.go:334] "Generic (PLEG): container finished" podID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerID="ac66a22531034d0affa1529759b983ac2f8598f419b9bc97a55b6b616367e33c" exitCode=0 Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.225390 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" event={"ID":"2e0e4181-e7fd-42fc-90b8-5e145f641943","Type":"ContainerDied","Data":"ac66a22531034d0affa1529759b983ac2f8598f419b9bc97a55b6b616367e33c"} Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.225590 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" event={"ID":"2e0e4181-e7fd-42fc-90b8-5e145f641943","Type":"ContainerDied","Data":"d9b490ca241d1556407e0500828829569d03aa8755918c9eb18666021b636743"} Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.225609 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9b490ca241d1556407e0500828829569d03aa8755918c9eb18666021b636743" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.252959 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.276806 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-dir\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.276854 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-serving-cert\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.276884 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-session\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.276915 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.276921 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-ocp-branding-template\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.276980 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-trusted-ca-bundle\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277008 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-login\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277661 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-router-certs\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277717 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-error\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277762 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-policies\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277785 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csz8f\" (UniqueName: \"kubernetes.io/projected/2e0e4181-e7fd-42fc-90b8-5e145f641943-kube-api-access-csz8f\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277778 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277810 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-service-ca\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277861 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-cliconfig\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277883 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-provider-selection\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.277922 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-idp-0-file-data\") pod \"2e0e4181-e7fd-42fc-90b8-5e145f641943\" (UID: \"2e0e4181-e7fd-42fc-90b8-5e145f641943\") " Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.278192 4829 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.278206 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.278249 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.278535 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.279186 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.282526 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.282697 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.283000 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.283837 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.285633 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.285734 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e0e4181-e7fd-42fc-90b8-5e145f641943-kube-api-access-csz8f" (OuterVolumeSpecName: "kube-api-access-csz8f") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "kube-api-access-csz8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.290406 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.291490 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.300072 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2e0e4181-e7fd-42fc-90b8-5e145f641943" (UID: "2e0e4181-e7fd-42fc-90b8-5e145f641943"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379839 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379886 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379906 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379920 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379936 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379948 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379962 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379975 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379986 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.379997 4829 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2e0e4181-e7fd-42fc-90b8-5e145f641943-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.380008 4829 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2e0e4181-e7fd-42fc-90b8-5e145f641943-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:44 crc kubenswrapper[4829]: I1002 07:20:44.380023 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csz8f\" (UniqueName: \"kubernetes.io/projected/2e0e4181-e7fd-42fc-90b8-5e145f641943-kube-api-access-csz8f\") on node \"crc\" DevicePath \"\"" Oct 02 07:20:45 crc kubenswrapper[4829]: I1002 07:20:45.233776 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6w49n" Oct 02 07:20:45 crc kubenswrapper[4829]: I1002 07:20:45.293766 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6w49n"] Oct 02 07:20:45 crc kubenswrapper[4829]: I1002 07:20:45.298801 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6w49n"] Oct 02 07:20:45 crc kubenswrapper[4829]: I1002 07:20:45.471926 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" path="/var/lib/kubelet/pods/2e0e4181-e7fd-42fc-90b8-5e145f641943/volumes" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.946867 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7559487fb5-wtm9s"] Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947588 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc175c8-2ee3-4e99-9ea2-650a50d81cfd" containerName="pruner" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947613 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc175c8-2ee3-4e99-9ea2-650a50d81cfd" containerName="pruner" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947633 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerName="oauth-openshift" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947645 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerName="oauth-openshift" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947660 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947673 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947693 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947704 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947719 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947731 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947745 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947757 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947772 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947784 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947803 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a759cf3a-59bf-478f-a87e-27165a2644e6" containerName="pruner" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947815 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a759cf3a-59bf-478f-a87e-27165a2644e6" containerName="pruner" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947829 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947840 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947857 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947870 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947888 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947900 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947922 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947934 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947952 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947963 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="extract-utilities" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.947977 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.947989 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: E1002 07:20:48.948006 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948019 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="extract-content" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948260 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d3333e7-33d0-4f71-b051-6980f83a77a4" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948286 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc175c8-2ee3-4e99-9ea2-650a50d81cfd" containerName="pruner" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948304 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a759cf3a-59bf-478f-a87e-27165a2644e6" containerName="pruner" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948337 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1f096a-3ff8-407c-87ef-d76e370aee43" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948363 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe403490-5471-4d5f-a032-29f0325f8ac3" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948391 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="330c4c1e-cb32-42fd-a3a8-8256339e5e17" containerName="registry-server" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.948415 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e0e4181-e7fd-42fc-90b8-5e145f641943" containerName="oauth-openshift" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.949059 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.954690 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.954775 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.955331 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.956034 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.956215 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.960496 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.960858 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.961129 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.961846 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.964094 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.964409 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.964848 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.966153 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7559487fb5-wtm9s"] Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.970571 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.973977 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 07:20:48 crc kubenswrapper[4829]: I1002 07:20:48.980932 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043488 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-audit-policies\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043532 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52k4p\" (UniqueName: \"kubernetes.io/projected/e3e847b6-20b2-4921-a9e3-83aff61817e4-kube-api-access-52k4p\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043587 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043670 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043728 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043779 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043843 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3e847b6-20b2-4921-a9e3-83aff61817e4-audit-dir\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043887 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043941 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043972 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.043993 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-login\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.044011 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-session\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.044028 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-error\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.044081 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.145189 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.145409 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.145520 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.145747 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-login\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.145939 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-session\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.146069 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-error\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.146295 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.146512 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-audit-policies\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.147004 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52k4p\" (UniqueName: \"kubernetes.io/projected/e3e847b6-20b2-4921-a9e3-83aff61817e4-kube-api-access-52k4p\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.147728 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.147994 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.148292 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.148730 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.148979 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3e847b6-20b2-4921-a9e3-83aff61817e4-audit-dir\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.149328 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.149330 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e3e847b6-20b2-4921-a9e3-83aff61817e4-audit-dir\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.148111 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-audit-policies\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.150572 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-service-ca\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.152144 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.153602 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-router-certs\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.153995 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-login\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.154129 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.156062 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-session\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.156837 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.156942 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-error\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.157107 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.158807 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e3e847b6-20b2-4921-a9e3-83aff61817e4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.170041 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52k4p\" (UniqueName: \"kubernetes.io/projected/e3e847b6-20b2-4921-a9e3-83aff61817e4-kube-api-access-52k4p\") pod \"oauth-openshift-7559487fb5-wtm9s\" (UID: \"e3e847b6-20b2-4921-a9e3-83aff61817e4\") " pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.285046 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:49 crc kubenswrapper[4829]: I1002 07:20:49.780768 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7559487fb5-wtm9s"] Oct 02 07:20:50 crc kubenswrapper[4829]: I1002 07:20:50.269692 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" event={"ID":"e3e847b6-20b2-4921-a9e3-83aff61817e4","Type":"ContainerStarted","Data":"6fbf8d53aed862b058ab124d66091484bf12a3c94008f4b50be34a49bdbf177c"} Oct 02 07:20:50 crc kubenswrapper[4829]: I1002 07:20:50.269742 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" event={"ID":"e3e847b6-20b2-4921-a9e3-83aff61817e4","Type":"ContainerStarted","Data":"ddd4c5f53540d5fe4eaef3cec8b4835e8f4ae616a14199076e70d0d42fe4a16f"} Oct 02 07:20:50 crc kubenswrapper[4829]: I1002 07:20:50.271262 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:20:50 crc kubenswrapper[4829]: I1002 07:20:50.290544 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" podStartSLOduration=32.290525134 podStartE2EDuration="32.290525134s" podCreationTimestamp="2025-10-02 07:20:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:20:50.289143462 +0000 UTC m=+241.628791917" watchObservedRunningTime="2025-10-02 07:20:50.290525134 +0000 UTC m=+241.630173539" Oct 02 07:20:50 crc kubenswrapper[4829]: I1002 07:20:50.530521 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7559487fb5-wtm9s" Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.727052 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xl9qq"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.728092 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xl9qq" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="registry-server" containerID="cri-o://ae6cdc959d27b6a3edaac0f0dff7d2a325452ba2786bf2ff4c052ba7874f2882" gracePeriod=30 Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.753133 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7rgt2"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.753489 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7rgt2" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="registry-server" containerID="cri-o://a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971" gracePeriod=30 Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.756602 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r77tn"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.756808 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" containerID="cri-o://8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939" gracePeriod=30 Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.762353 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cm84"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.762560 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7cm84" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="registry-server" containerID="cri-o://7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c" gracePeriod=30 Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.765934 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hj5v"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.766297 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6hj5v" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="registry-server" containerID="cri-o://c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad" gracePeriod=30 Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.778461 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbdf8"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.779315 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.789890 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbdf8"] Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.959889 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.960198 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:09 crc kubenswrapper[4829]: I1002 07:21:09.960260 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zqm\" (UniqueName: \"kubernetes.io/projected/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-kube-api-access-c6zqm\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.061804 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.061884 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.061926 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zqm\" (UniqueName: \"kubernetes.io/projected/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-kube-api-access-c6zqm\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.065477 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.069214 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.088662 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zqm\" (UniqueName: \"kubernetes.io/projected/156b7e29-6cb2-4979-9ef3-c9f7904ecc0f-kube-api-access-c6zqm\") pod \"marketplace-operator-79b997595-bbdf8\" (UID: \"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.157611 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.211172 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.216656 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.248217 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365277 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc62w\" (UniqueName: \"kubernetes.io/projected/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-kube-api-access-fc62w\") pod \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365572 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-utilities\") pod \"0357590d-1dad-4ef3-b422-8b25a71953ab\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365621 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-catalog-content\") pod \"0357590d-1dad-4ef3-b422-8b25a71953ab\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365658 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-operator-metrics\") pod \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365685 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-trusted-ca\") pod \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\" (UID: \"72e0891e-a4b5-43c2-918f-e5ffdc53fe20\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365721 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv4s2\" (UniqueName: \"kubernetes.io/projected/37c0917f-a686-4dc9-99d3-28ff284ca2e0-kube-api-access-fv4s2\") pod \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365774 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gb9kg\" (UniqueName: \"kubernetes.io/projected/0357590d-1dad-4ef3-b422-8b25a71953ab-kube-api-access-gb9kg\") pod \"0357590d-1dad-4ef3-b422-8b25a71953ab\" (UID: \"0357590d-1dad-4ef3-b422-8b25a71953ab\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365810 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-utilities\") pod \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.365833 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-catalog-content\") pod \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\" (UID: \"37c0917f-a686-4dc9-99d3-28ff284ca2e0\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.369365 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-utilities" (OuterVolumeSpecName: "utilities") pod "0357590d-1dad-4ef3-b422-8b25a71953ab" (UID: "0357590d-1dad-4ef3-b422-8b25a71953ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.370331 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-utilities" (OuterVolumeSpecName: "utilities") pod "37c0917f-a686-4dc9-99d3-28ff284ca2e0" (UID: "37c0917f-a686-4dc9-99d3-28ff284ca2e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.371067 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "72e0891e-a4b5-43c2-918f-e5ffdc53fe20" (UID: "72e0891e-a4b5-43c2-918f-e5ffdc53fe20"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.371742 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0357590d-1dad-4ef3-b422-8b25a71953ab-kube-api-access-gb9kg" (OuterVolumeSpecName: "kube-api-access-gb9kg") pod "0357590d-1dad-4ef3-b422-8b25a71953ab" (UID: "0357590d-1dad-4ef3-b422-8b25a71953ab"). InnerVolumeSpecName "kube-api-access-gb9kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.372072 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-kube-api-access-fc62w" (OuterVolumeSpecName: "kube-api-access-fc62w") pod "72e0891e-a4b5-43c2-918f-e5ffdc53fe20" (UID: "72e0891e-a4b5-43c2-918f-e5ffdc53fe20"). InnerVolumeSpecName "kube-api-access-fc62w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.372183 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c0917f-a686-4dc9-99d3-28ff284ca2e0-kube-api-access-fv4s2" (OuterVolumeSpecName: "kube-api-access-fv4s2") pod "37c0917f-a686-4dc9-99d3-28ff284ca2e0" (UID: "37c0917f-a686-4dc9-99d3-28ff284ca2e0"). InnerVolumeSpecName "kube-api-access-fv4s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.372655 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "72e0891e-a4b5-43c2-918f-e5ffdc53fe20" (UID: "72e0891e-a4b5-43c2-918f-e5ffdc53fe20"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.386546 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37c0917f-a686-4dc9-99d3-28ff284ca2e0" (UID: "37c0917f-a686-4dc9-99d3-28ff284ca2e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.395428 4829 generic.go:334] "Generic (PLEG): container finished" podID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerID="8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939" exitCode=0 Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.395509 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" event={"ID":"72e0891e-a4b5-43c2-918f-e5ffdc53fe20","Type":"ContainerDied","Data":"8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.395540 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" event={"ID":"72e0891e-a4b5-43c2-918f-e5ffdc53fe20","Type":"ContainerDied","Data":"fda293488888519a85d3094cd9573cc5bc8dd047d59a3e8db25b46a428d7e9d2"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.395556 4829 scope.go:117] "RemoveContainer" containerID="8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.395663 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-r77tn" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.400056 4829 generic.go:334] "Generic (PLEG): container finished" podID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerID="ae6cdc959d27b6a3edaac0f0dff7d2a325452ba2786bf2ff4c052ba7874f2882" exitCode=0 Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.400120 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl9qq" event={"ID":"3cd1d972-be99-4ae8-b965-dd55a308e3a9","Type":"ContainerDied","Data":"ae6cdc959d27b6a3edaac0f0dff7d2a325452ba2786bf2ff4c052ba7874f2882"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.410751 4829 generic.go:334] "Generic (PLEG): container finished" podID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerID="c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad" exitCode=0 Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.410851 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hj5v" event={"ID":"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55","Type":"ContainerDied","Data":"c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.414030 4829 generic.go:334] "Generic (PLEG): container finished" podID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerID="a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971" exitCode=0 Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.414062 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rgt2" event={"ID":"0357590d-1dad-4ef3-b422-8b25a71953ab","Type":"ContainerDied","Data":"a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.414089 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7rgt2" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.414114 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7rgt2" event={"ID":"0357590d-1dad-4ef3-b422-8b25a71953ab","Type":"ContainerDied","Data":"5bf331f22acbd276e8317f5908689df5c15159c34ce5285e336ae709584d232c"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.416344 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0357590d-1dad-4ef3-b422-8b25a71953ab" (UID: "0357590d-1dad-4ef3-b422-8b25a71953ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.417104 4829 generic.go:334] "Generic (PLEG): container finished" podID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerID="7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c" exitCode=0 Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.417141 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cm84" event={"ID":"37c0917f-a686-4dc9-99d3-28ff284ca2e0","Type":"ContainerDied","Data":"7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.417167 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7cm84" event={"ID":"37c0917f-a686-4dc9-99d3-28ff284ca2e0","Type":"ContainerDied","Data":"786eaed23ca90890286d08435499c33a93d64775a4934dc452e727ff0e5a7f43"} Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.417198 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7cm84" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.432735 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r77tn"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.432779 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-r77tn"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467401 4829 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467425 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv4s2\" (UniqueName: \"kubernetes.io/projected/37c0917f-a686-4dc9-99d3-28ff284ca2e0-kube-api-access-fv4s2\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467435 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gb9kg\" (UniqueName: \"kubernetes.io/projected/0357590d-1dad-4ef3-b422-8b25a71953ab-kube-api-access-gb9kg\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467444 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467454 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c0917f-a686-4dc9-99d3-28ff284ca2e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467464 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc62w\" (UniqueName: \"kubernetes.io/projected/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-kube-api-access-fc62w\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467472 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467480 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0357590d-1dad-4ef3-b422-8b25a71953ab-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.467488 4829 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72e0891e-a4b5-43c2-918f-e5ffdc53fe20-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.482804 4829 scope.go:117] "RemoveContainer" containerID="8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.483420 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939\": container with ID starting with 8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939 not found: ID does not exist" containerID="8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.483449 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939"} err="failed to get container status \"8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939\": rpc error: code = NotFound desc = could not find container \"8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939\": container with ID starting with 8a1a0219de0ae98c9c49f701c53fc2f271ea9cd5f21ca1d43eff84cb511b5939 not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.483468 4829 scope.go:117] "RemoveContainer" containerID="a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.497733 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cm84"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.504657 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7cm84"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.509360 4829 scope.go:117] "RemoveContainer" containerID="92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.545603 4829 scope.go:117] "RemoveContainer" containerID="54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.572632 4829 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad is running failed: container process not found" containerID="c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.572998 4829 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad is running failed: container process not found" containerID="c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.573276 4829 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad is running failed: container process not found" containerID="c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.573320 4829 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-6hj5v" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="registry-server" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.573799 4829 scope.go:117] "RemoveContainer" containerID="a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.574393 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971\": container with ID starting with a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971 not found: ID does not exist" containerID="a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.574418 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971"} err="failed to get container status \"a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971\": rpc error: code = NotFound desc = could not find container \"a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971\": container with ID starting with a26ddb9d6df5927e2c0f4f37cb024dc38d2b92bcc3104ba2c38e36ba55107971 not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.574436 4829 scope.go:117] "RemoveContainer" containerID="92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.574677 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341\": container with ID starting with 92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341 not found: ID does not exist" containerID="92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.574721 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341"} err="failed to get container status \"92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341\": rpc error: code = NotFound desc = could not find container \"92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341\": container with ID starting with 92460c11b466a99ca8b5524cac3958d939141d8d27400d8c27e64517a282d341 not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.574746 4829 scope.go:117] "RemoveContainer" containerID="54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.575148 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c\": container with ID starting with 54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c not found: ID does not exist" containerID="54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.575189 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c"} err="failed to get container status \"54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c\": rpc error: code = NotFound desc = could not find container \"54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c\": container with ID starting with 54f4fffd27afe243cba36d5f231a9d4b38361165bef8aa83fcdf074bf2bfa07c not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.575216 4829 scope.go:117] "RemoveContainer" containerID="7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.575635 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bbdf8"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.592432 4829 scope.go:117] "RemoveContainer" containerID="a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.622022 4829 scope.go:117] "RemoveContainer" containerID="709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.644047 4829 scope.go:117] "RemoveContainer" containerID="7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.644466 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c\": container with ID starting with 7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c not found: ID does not exist" containerID="7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.644519 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c"} err="failed to get container status \"7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c\": rpc error: code = NotFound desc = could not find container \"7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c\": container with ID starting with 7075c52117db3d68f1a9aa102e98b6b4a780397ac0f5adf3a393b0e72363946c not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.644546 4829 scope.go:117] "RemoveContainer" containerID="a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.645177 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b\": container with ID starting with a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b not found: ID does not exist" containerID="a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.645209 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b"} err="failed to get container status \"a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b\": rpc error: code = NotFound desc = could not find container \"a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b\": container with ID starting with a4586d48aab9b12edceb7f0d845953a9aa0b7b382b7150e69a9994ff1a1cd36b not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.645239 4829 scope.go:117] "RemoveContainer" containerID="709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8" Oct 02 07:21:10 crc kubenswrapper[4829]: E1002 07:21:10.645514 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8\": container with ID starting with 709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8 not found: ID does not exist" containerID="709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.645530 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8"} err="failed to get container status \"709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8\": rpc error: code = NotFound desc = could not find container \"709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8\": container with ID starting with 709c0bf942d924d87176edfbb76ab5fe87e64c6e3aed230685e3f38aa61ac3e8 not found: ID does not exist" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.655819 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.726318 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.775126 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-catalog-content\") pod \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.775235 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-utilities\") pod \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.775268 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnq4t\" (UniqueName: \"kubernetes.io/projected/3cd1d972-be99-4ae8-b965-dd55a308e3a9-kube-api-access-fnq4t\") pod \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\" (UID: \"3cd1d972-be99-4ae8-b965-dd55a308e3a9\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.775383 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7rgt2"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.775919 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-utilities" (OuterVolumeSpecName: "utilities") pod "3cd1d972-be99-4ae8-b965-dd55a308e3a9" (UID: "3cd1d972-be99-4ae8-b965-dd55a308e3a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.777961 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd1d972-be99-4ae8-b965-dd55a308e3a9-kube-api-access-fnq4t" (OuterVolumeSpecName: "kube-api-access-fnq4t") pod "3cd1d972-be99-4ae8-b965-dd55a308e3a9" (UID: "3cd1d972-be99-4ae8-b965-dd55a308e3a9"). InnerVolumeSpecName "kube-api-access-fnq4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.778131 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7rgt2"] Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.819103 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cd1d972-be99-4ae8-b965-dd55a308e3a9" (UID: "3cd1d972-be99-4ae8-b965-dd55a308e3a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.876600 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-catalog-content\") pod \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.876656 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-utilities\") pod \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.876712 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5598\" (UniqueName: \"kubernetes.io/projected/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-kube-api-access-z5598\") pod \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\" (UID: \"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55\") " Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.876929 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.876945 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd1d972-be99-4ae8-b965-dd55a308e3a9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.876954 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnq4t\" (UniqueName: \"kubernetes.io/projected/3cd1d972-be99-4ae8-b965-dd55a308e3a9-kube-api-access-fnq4t\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.878328 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-utilities" (OuterVolumeSpecName: "utilities") pod "f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" (UID: "f210ccef-7e77-4e8c-97ca-d2e3abeeaa55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.880779 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-kube-api-access-z5598" (OuterVolumeSpecName: "kube-api-access-z5598") pod "f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" (UID: "f210ccef-7e77-4e8c-97ca-d2e3abeeaa55"). InnerVolumeSpecName "kube-api-access-z5598". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.970669 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" (UID: "f210ccef-7e77-4e8c-97ca-d2e3abeeaa55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.977489 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.977515 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5598\" (UniqueName: \"kubernetes.io/projected/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-kube-api-access-z5598\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:10 crc kubenswrapper[4829]: I1002 07:21:10.977525 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.423413 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xl9qq" event={"ID":"3cd1d972-be99-4ae8-b965-dd55a308e3a9","Type":"ContainerDied","Data":"64b576ad586a8f4e030892b0b0502accd8cad58593ecbf288ed9a40fbf3b338f"} Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.423780 4829 scope.go:117] "RemoveContainer" containerID="ae6cdc959d27b6a3edaac0f0dff7d2a325452ba2786bf2ff4c052ba7874f2882" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.423456 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xl9qq" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.425713 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6hj5v" event={"ID":"f210ccef-7e77-4e8c-97ca-d2e3abeeaa55","Type":"ContainerDied","Data":"353c270ee08312529e310e96e4fe4e4d254fbd318c728db10f122cc03c77a8be"} Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.425753 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6hj5v" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.430266 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" event={"ID":"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f","Type":"ContainerStarted","Data":"940d8385c26040ac6d8762edd0e5ec44cf87f267723a9225cbbd950ce79ea3f0"} Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.430299 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" event={"ID":"156b7e29-6cb2-4979-9ef3-c9f7904ecc0f","Type":"ContainerStarted","Data":"b493bcf6d309a36fb18e67c8229b2375009a9fddfe0b97416f75936d7ff67c99"} Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.430558 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.434640 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.449482 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bbdf8" podStartSLOduration=2.449453946 podStartE2EDuration="2.449453946s" podCreationTimestamp="2025-10-02 07:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:21:11.444424479 +0000 UTC m=+262.784072924" watchObservedRunningTime="2025-10-02 07:21:11.449453946 +0000 UTC m=+262.789102391" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.450654 4829 scope.go:117] "RemoveContainer" containerID="3af68e7aedcff8a48f0ba5232eba7ed9aa453f030a85578b31eb062b36997827" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.474695 4829 scope.go:117] "RemoveContainer" containerID="ae33fa9287d69a7524a09ec08cc9a5d39c61ec3fb2fa7d4936cb372c26499f70" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.495788 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" path="/var/lib/kubelet/pods/0357590d-1dad-4ef3-b422-8b25a71953ab/volumes" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.498017 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" path="/var/lib/kubelet/pods/37c0917f-a686-4dc9-99d3-28ff284ca2e0/volumes" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.501248 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" path="/var/lib/kubelet/pods/72e0891e-a4b5-43c2-918f-e5ffdc53fe20/volumes" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.501837 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6hj5v"] Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.501864 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6hj5v"] Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.504881 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xl9qq"] Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.506778 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xl9qq"] Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.520505 4829 scope.go:117] "RemoveContainer" containerID="c5c2ffc05de31a88ec5d19d45676d48f00bbb3a1b49f457d23b3726054e878ad" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.533886 4829 scope.go:117] "RemoveContainer" containerID="7680fe23e4b068b7198564f6aff3fd576d57eb20a8e229ff1c6134e8f4f86cf4" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.550874 4829 scope.go:117] "RemoveContainer" containerID="1be1992547c15e9708e6605b74f491927916ad5e1a129016ffbc81d2b9a3d873" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.943371 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9spqb"] Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.943895 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.943909 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.943923 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.943930 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.943945 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.943953 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.943963 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.943971 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.943980 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.943988 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.943999 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944006 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944014 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944021 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944029 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944037 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944048 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944055 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944068 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944075 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="extract-content" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944086 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944094 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944102 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944109 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: E1002 07:21:11.944119 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944127 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="extract-utilities" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944263 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944276 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c0917f-a686-4dc9-99d3-28ff284ca2e0" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944290 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0357590d-1dad-4ef3-b422-8b25a71953ab" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944303 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" containerName="registry-server" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.944312 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e0891e-a4b5-43c2-918f-e5ffdc53fe20" containerName="marketplace-operator" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.945023 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.948018 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 07:21:11 crc kubenswrapper[4829]: I1002 07:21:11.960762 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9spqb"] Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.090369 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-utilities\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.090444 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-catalog-content\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.090517 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4wxt\" (UniqueName: \"kubernetes.io/projected/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-kube-api-access-w4wxt\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.151395 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pc6x5"] Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.154882 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.157975 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.162095 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pc6x5"] Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.191456 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4wxt\" (UniqueName: \"kubernetes.io/projected/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-kube-api-access-w4wxt\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.191641 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-utilities\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.197700 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-utilities\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.197906 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-catalog-content\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.198898 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-catalog-content\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.229853 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4wxt\" (UniqueName: \"kubernetes.io/projected/cef47a6d-4a19-45e0-9dca-2ab3a4db27dc-kube-api-access-w4wxt\") pod \"redhat-marketplace-9spqb\" (UID: \"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc\") " pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.279514 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.299136 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64a03ff-edef-475b-ae70-739d43cefd48-utilities\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.299183 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8lkp\" (UniqueName: \"kubernetes.io/projected/b64a03ff-edef-475b-ae70-739d43cefd48-kube-api-access-q8lkp\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.299405 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64a03ff-edef-475b-ae70-739d43cefd48-catalog-content\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.401059 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64a03ff-edef-475b-ae70-739d43cefd48-utilities\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.401328 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8lkp\" (UniqueName: \"kubernetes.io/projected/b64a03ff-edef-475b-ae70-739d43cefd48-kube-api-access-q8lkp\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.401402 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64a03ff-edef-475b-ae70-739d43cefd48-catalog-content\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.401933 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b64a03ff-edef-475b-ae70-739d43cefd48-catalog-content\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.402168 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b64a03ff-edef-475b-ae70-739d43cefd48-utilities\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.424074 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8lkp\" (UniqueName: \"kubernetes.io/projected/b64a03ff-edef-475b-ae70-739d43cefd48-kube-api-access-q8lkp\") pod \"redhat-operators-pc6x5\" (UID: \"b64a03ff-edef-475b-ae70-739d43cefd48\") " pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.473888 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9spqb"] Oct 02 07:21:12 crc kubenswrapper[4829]: W1002 07:21:12.480474 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcef47a6d_4a19_45e0_9dca_2ab3a4db27dc.slice/crio-926cd016f66cf67faf0a3a13a949d58b8e3436c0f8b79f908950c49695cf1375 WatchSource:0}: Error finding container 926cd016f66cf67faf0a3a13a949d58b8e3436c0f8b79f908950c49695cf1375: Status 404 returned error can't find the container with id 926cd016f66cf67faf0a3a13a949d58b8e3436c0f8b79f908950c49695cf1375 Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.501115 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:12 crc kubenswrapper[4829]: I1002 07:21:12.900032 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pc6x5"] Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.451421 4829 generic.go:334] "Generic (PLEG): container finished" podID="b64a03ff-edef-475b-ae70-739d43cefd48" containerID="92926a24680df4801c9ad47aeadc090173f7d29e82a6e3870a5670792e1c7837" exitCode=0 Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.451489 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pc6x5" event={"ID":"b64a03ff-edef-475b-ae70-739d43cefd48","Type":"ContainerDied","Data":"92926a24680df4801c9ad47aeadc090173f7d29e82a6e3870a5670792e1c7837"} Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.451516 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pc6x5" event={"ID":"b64a03ff-edef-475b-ae70-739d43cefd48","Type":"ContainerStarted","Data":"fb03e8fb83fb7d1156c185d59fb7b898d2e9d72dd38573f571ad340ac9d080eb"} Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.455508 4829 generic.go:334] "Generic (PLEG): container finished" podID="cef47a6d-4a19-45e0-9dca-2ab3a4db27dc" containerID="cde34337f58e2827da01b62757f7ecddb8b0f3d0298d5281dddcb7720ed7c271" exitCode=0 Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.456146 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9spqb" event={"ID":"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc","Type":"ContainerDied","Data":"cde34337f58e2827da01b62757f7ecddb8b0f3d0298d5281dddcb7720ed7c271"} Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.456174 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9spqb" event={"ID":"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc","Type":"ContainerStarted","Data":"926cd016f66cf67faf0a3a13a949d58b8e3436c0f8b79f908950c49695cf1375"} Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.474755 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd1d972-be99-4ae8-b965-dd55a308e3a9" path="/var/lib/kubelet/pods/3cd1d972-be99-4ae8-b965-dd55a308e3a9/volumes" Oct 02 07:21:13 crc kubenswrapper[4829]: I1002 07:21:13.481999 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f210ccef-7e77-4e8c-97ca-d2e3abeeaa55" path="/var/lib/kubelet/pods/f210ccef-7e77-4e8c-97ca-d2e3abeeaa55/volumes" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.343569 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j9jpj"] Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.344985 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.347636 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.359820 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9jpj"] Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.461267 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pc6x5" event={"ID":"b64a03ff-edef-475b-ae70-739d43cefd48","Type":"ContainerStarted","Data":"995cc1beb3744973ff6c0b8c4112f5ea37871a639f5d3af4836ed447c9afcdbe"} Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.466883 4829 generic.go:334] "Generic (PLEG): container finished" podID="cef47a6d-4a19-45e0-9dca-2ab3a4db27dc" containerID="bf3bf547e9c8698180829c30c54f390002d49ab95a871af5405cbe947febf953" exitCode=0 Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.466939 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9spqb" event={"ID":"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc","Type":"ContainerDied","Data":"bf3bf547e9c8698180829c30c54f390002d49ab95a871af5405cbe947febf953"} Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.525909 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-catalog-content\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.525981 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-utilities\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.526017 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptdw4\" (UniqueName: \"kubernetes.io/projected/b3d721af-feab-49f1-833f-fcf222170ce3-kube-api-access-ptdw4\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.546255 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kfj6n"] Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.547098 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.549830 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.563459 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfj6n"] Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.627950 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-catalog-content\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.628029 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-utilities\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.628048 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptdw4\" (UniqueName: \"kubernetes.io/projected/b3d721af-feab-49f1-833f-fcf222170ce3-kube-api-access-ptdw4\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.628552 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-utilities\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.628741 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-catalog-content\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.656882 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptdw4\" (UniqueName: \"kubernetes.io/projected/b3d721af-feab-49f1-833f-fcf222170ce3-kube-api-access-ptdw4\") pod \"community-operators-j9jpj\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.666298 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.729105 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-catalog-content\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.729585 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-utilities\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.729712 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cjl7\" (UniqueName: \"kubernetes.io/projected/f8d76d5c-3014-461e-8bac-253a85703dd9-kube-api-access-2cjl7\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.831260 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-utilities\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.831328 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cjl7\" (UniqueName: \"kubernetes.io/projected/f8d76d5c-3014-461e-8bac-253a85703dd9-kube-api-access-2cjl7\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.831390 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-catalog-content\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.831970 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-catalog-content\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.832292 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-utilities\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.853873 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cjl7\" (UniqueName: \"kubernetes.io/projected/f8d76d5c-3014-461e-8bac-253a85703dd9-kube-api-access-2cjl7\") pod \"certified-operators-kfj6n\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:14 crc kubenswrapper[4829]: I1002 07:21:14.862692 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.039576 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kfj6n"] Oct 02 07:21:15 crc kubenswrapper[4829]: W1002 07:21:15.068693 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8d76d5c_3014_461e_8bac_253a85703dd9.slice/crio-2e8d028765e7c4b21628d56ad9314d70cce671f1a165644accc2e603290b5ead WatchSource:0}: Error finding container 2e8d028765e7c4b21628d56ad9314d70cce671f1a165644accc2e603290b5ead: Status 404 returned error can't find the container with id 2e8d028765e7c4b21628d56ad9314d70cce671f1a165644accc2e603290b5ead Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.072991 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j9jpj"] Oct 02 07:21:15 crc kubenswrapper[4829]: W1002 07:21:15.076925 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3d721af_feab_49f1_833f_fcf222170ce3.slice/crio-a2a736fce4d1b8cbf9a532317d10dc95f60ce5231f7ac953b3ea15a5db3a2449 WatchSource:0}: Error finding container a2a736fce4d1b8cbf9a532317d10dc95f60ce5231f7ac953b3ea15a5db3a2449: Status 404 returned error can't find the container with id a2a736fce4d1b8cbf9a532317d10dc95f60ce5231f7ac953b3ea15a5db3a2449 Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.474175 4829 generic.go:334] "Generic (PLEG): container finished" podID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerID="c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132" exitCode=0 Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.474533 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfj6n" event={"ID":"f8d76d5c-3014-461e-8bac-253a85703dd9","Type":"ContainerDied","Data":"c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132"} Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.474562 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfj6n" event={"ID":"f8d76d5c-3014-461e-8bac-253a85703dd9","Type":"ContainerStarted","Data":"2e8d028765e7c4b21628d56ad9314d70cce671f1a165644accc2e603290b5ead"} Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.478170 4829 generic.go:334] "Generic (PLEG): container finished" podID="b64a03ff-edef-475b-ae70-739d43cefd48" containerID="995cc1beb3744973ff6c0b8c4112f5ea37871a639f5d3af4836ed447c9afcdbe" exitCode=0 Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.478258 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pc6x5" event={"ID":"b64a03ff-edef-475b-ae70-739d43cefd48","Type":"ContainerDied","Data":"995cc1beb3744973ff6c0b8c4112f5ea37871a639f5d3af4836ed447c9afcdbe"} Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.481142 4829 generic.go:334] "Generic (PLEG): container finished" podID="b3d721af-feab-49f1-833f-fcf222170ce3" containerID="ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2" exitCode=0 Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.481223 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9jpj" event={"ID":"b3d721af-feab-49f1-833f-fcf222170ce3","Type":"ContainerDied","Data":"ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2"} Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.481277 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9jpj" event={"ID":"b3d721af-feab-49f1-833f-fcf222170ce3","Type":"ContainerStarted","Data":"a2a736fce4d1b8cbf9a532317d10dc95f60ce5231f7ac953b3ea15a5db3a2449"} Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.484415 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9spqb" event={"ID":"cef47a6d-4a19-45e0-9dca-2ab3a4db27dc","Type":"ContainerStarted","Data":"02b14f6051442c98b1ae1632b8a9a4809c4dc0c130c7da4f422ac0a13f0d6175"} Oct 02 07:21:15 crc kubenswrapper[4829]: I1002 07:21:15.516999 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9spqb" podStartSLOduration=2.999067954 podStartE2EDuration="4.516980828s" podCreationTimestamp="2025-10-02 07:21:11 +0000 UTC" firstStartedPulling="2025-10-02 07:21:13.458983107 +0000 UTC m=+264.798631512" lastFinishedPulling="2025-10-02 07:21:14.976895991 +0000 UTC m=+266.316544386" observedRunningTime="2025-10-02 07:21:15.516296787 +0000 UTC m=+266.855945192" watchObservedRunningTime="2025-10-02 07:21:15.516980828 +0000 UTC m=+266.856629233" Oct 02 07:21:16 crc kubenswrapper[4829]: I1002 07:21:16.491805 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pc6x5" event={"ID":"b64a03ff-edef-475b-ae70-739d43cefd48","Type":"ContainerStarted","Data":"4546153bda644ae99c42425b7dbcf0a7d7c45c583a3bacff018932d09ce00332"} Oct 02 07:21:16 crc kubenswrapper[4829]: I1002 07:21:16.493482 4829 generic.go:334] "Generic (PLEG): container finished" podID="b3d721af-feab-49f1-833f-fcf222170ce3" containerID="43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3" exitCode=0 Oct 02 07:21:16 crc kubenswrapper[4829]: I1002 07:21:16.493566 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9jpj" event={"ID":"b3d721af-feab-49f1-833f-fcf222170ce3","Type":"ContainerDied","Data":"43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3"} Oct 02 07:21:16 crc kubenswrapper[4829]: I1002 07:21:16.495394 4829 generic.go:334] "Generic (PLEG): container finished" podID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerID="1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380" exitCode=0 Oct 02 07:21:16 crc kubenswrapper[4829]: I1002 07:21:16.495536 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfj6n" event={"ID":"f8d76d5c-3014-461e-8bac-253a85703dd9","Type":"ContainerDied","Data":"1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380"} Oct 02 07:21:16 crc kubenswrapper[4829]: I1002 07:21:16.511843 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pc6x5" podStartSLOduration=2.03964469 podStartE2EDuration="4.511829406s" podCreationTimestamp="2025-10-02 07:21:12 +0000 UTC" firstStartedPulling="2025-10-02 07:21:13.454731816 +0000 UTC m=+264.794380251" lastFinishedPulling="2025-10-02 07:21:15.926916562 +0000 UTC m=+267.266564967" observedRunningTime="2025-10-02 07:21:16.509176163 +0000 UTC m=+267.848824578" watchObservedRunningTime="2025-10-02 07:21:16.511829406 +0000 UTC m=+267.851477811" Oct 02 07:21:17 crc kubenswrapper[4829]: I1002 07:21:17.502268 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9jpj" event={"ID":"b3d721af-feab-49f1-833f-fcf222170ce3","Type":"ContainerStarted","Data":"8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad"} Oct 02 07:21:17 crc kubenswrapper[4829]: I1002 07:21:17.505730 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfj6n" event={"ID":"f8d76d5c-3014-461e-8bac-253a85703dd9","Type":"ContainerStarted","Data":"94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698"} Oct 02 07:21:17 crc kubenswrapper[4829]: I1002 07:21:17.523552 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j9jpj" podStartSLOduration=2.059584304 podStartE2EDuration="3.523536857s" podCreationTimestamp="2025-10-02 07:21:14 +0000 UTC" firstStartedPulling="2025-10-02 07:21:15.483904568 +0000 UTC m=+266.823552983" lastFinishedPulling="2025-10-02 07:21:16.947857121 +0000 UTC m=+268.287505536" observedRunningTime="2025-10-02 07:21:17.521872355 +0000 UTC m=+268.861520760" watchObservedRunningTime="2025-10-02 07:21:17.523536857 +0000 UTC m=+268.863185262" Oct 02 07:21:17 crc kubenswrapper[4829]: I1002 07:21:17.547460 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kfj6n" podStartSLOduration=2.1413746 podStartE2EDuration="3.547442641s" podCreationTimestamp="2025-10-02 07:21:14 +0000 UTC" firstStartedPulling="2025-10-02 07:21:15.476022683 +0000 UTC m=+266.815671088" lastFinishedPulling="2025-10-02 07:21:16.882090724 +0000 UTC m=+268.221739129" observedRunningTime="2025-10-02 07:21:17.539602928 +0000 UTC m=+268.879251333" watchObservedRunningTime="2025-10-02 07:21:17.547442641 +0000 UTC m=+268.887091046" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.280207 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.280971 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.353742 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.501891 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.501932 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.547747 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.574794 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9spqb" Oct 02 07:21:22 crc kubenswrapper[4829]: I1002 07:21:22.587784 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pc6x5" Oct 02 07:21:24 crc kubenswrapper[4829]: I1002 07:21:24.667148 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:24 crc kubenswrapper[4829]: I1002 07:21:24.667657 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:24 crc kubenswrapper[4829]: I1002 07:21:24.704751 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:21:24 crc kubenswrapper[4829]: I1002 07:21:24.863375 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:24 crc kubenswrapper[4829]: I1002 07:21:24.863448 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:24 crc kubenswrapper[4829]: I1002 07:21:24.921826 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:25 crc kubenswrapper[4829]: I1002 07:21:25.586868 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:21:25 crc kubenswrapper[4829]: I1002 07:21:25.591298 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:22:55 crc kubenswrapper[4829]: I1002 07:22:55.329659 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:22:55 crc kubenswrapper[4829]: I1002 07:22:55.330298 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:23:25 crc kubenswrapper[4829]: I1002 07:23:25.329379 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:23:25 crc kubenswrapper[4829]: I1002 07:23:25.329984 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.329336 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.329951 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.330022 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.330889 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eaf4b9893a38009f8a714f08d03494fe5d6ecc601b0a3a024d68f3487caee9a2"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.330991 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://eaf4b9893a38009f8a714f08d03494fe5d6ecc601b0a3a024d68f3487caee9a2" gracePeriod=600 Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.517197 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="eaf4b9893a38009f8a714f08d03494fe5d6ecc601b0a3a024d68f3487caee9a2" exitCode=0 Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.517359 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"eaf4b9893a38009f8a714f08d03494fe5d6ecc601b0a3a024d68f3487caee9a2"} Oct 02 07:23:55 crc kubenswrapper[4829]: I1002 07:23:55.517657 4829 scope.go:117] "RemoveContainer" containerID="fe10048d833b2283dbe569b679bd4232ea1719dab9f72413d8d575f0cfbdc146" Oct 02 07:23:56 crc kubenswrapper[4829]: I1002 07:23:56.529658 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"3c344e588de3b2b45a68b8eaf305327aebb97a91305e92834159bf1c925efb2c"} Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.631442 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhmv6"] Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.632690 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.684721 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhmv6"] Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731303 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24c0b69a-38c9-4082-aa39-cda8997b1144-trusted-ca\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731682 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jksf\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-kube-api-access-6jksf\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731719 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-registry-tls\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731748 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24c0b69a-38c9-4082-aa39-cda8997b1144-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731885 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731946 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24c0b69a-38c9-4082-aa39-cda8997b1144-registry-certificates\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.731989 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24c0b69a-38c9-4082-aa39-cda8997b1144-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.732050 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-bound-sa-token\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.761152 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833359 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24c0b69a-38c9-4082-aa39-cda8997b1144-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833435 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24c0b69a-38c9-4082-aa39-cda8997b1144-registry-certificates\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833497 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-bound-sa-token\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833556 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24c0b69a-38c9-4082-aa39-cda8997b1144-trusted-ca\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833588 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jksf\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-kube-api-access-6jksf\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833632 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-registry-tls\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.833670 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24c0b69a-38c9-4082-aa39-cda8997b1144-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.834947 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/24c0b69a-38c9-4082-aa39-cda8997b1144-ca-trust-extracted\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.835281 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24c0b69a-38c9-4082-aa39-cda8997b1144-trusted-ca\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.835893 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/24c0b69a-38c9-4082-aa39-cda8997b1144-registry-certificates\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.845686 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/24c0b69a-38c9-4082-aa39-cda8997b1144-installation-pull-secrets\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.845881 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-registry-tls\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.852760 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jksf\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-kube-api-access-6jksf\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.856551 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24c0b69a-38c9-4082-aa39-cda8997b1144-bound-sa-token\") pod \"image-registry-66df7c8f76-hhmv6\" (UID: \"24c0b69a-38c9-4082-aa39-cda8997b1144\") " pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:34 crc kubenswrapper[4829]: I1002 07:24:34.953120 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:35 crc kubenswrapper[4829]: I1002 07:24:35.438055 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-hhmv6"] Oct 02 07:24:35 crc kubenswrapper[4829]: I1002 07:24:35.786987 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" event={"ID":"24c0b69a-38c9-4082-aa39-cda8997b1144","Type":"ContainerStarted","Data":"b7a7ab164103132d8b907335e12903e27aef45113747917e3cb06de986d8ba96"} Oct 02 07:24:35 crc kubenswrapper[4829]: I1002 07:24:35.787038 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" event={"ID":"24c0b69a-38c9-4082-aa39-cda8997b1144","Type":"ContainerStarted","Data":"7f197223b730ec8204d176987a59021bccfe3b17e5f63a75c4a37d7057de162d"} Oct 02 07:24:35 crc kubenswrapper[4829]: I1002 07:24:35.787158 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:35 crc kubenswrapper[4829]: I1002 07:24:35.812713 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" podStartSLOduration=1.812682874 podStartE2EDuration="1.812682874s" podCreationTimestamp="2025-10-02 07:24:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:24:35.805487446 +0000 UTC m=+467.145135871" watchObservedRunningTime="2025-10-02 07:24:35.812682874 +0000 UTC m=+467.152331319" Oct 02 07:24:54 crc kubenswrapper[4829]: I1002 07:24:54.964480 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-hhmv6" Oct 02 07:24:55 crc kubenswrapper[4829]: I1002 07:24:55.022353 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qx6xb"] Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.057466 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" podUID="9c41ceaa-bc0a-4576-b309-0e7679130b71" containerName="registry" containerID="cri-o://b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c" gracePeriod=30 Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.430579 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.577906 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-certificates\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.577998 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-bound-sa-token\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.578044 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmm5k\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-kube-api-access-tmm5k\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.578081 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c41ceaa-bc0a-4576-b309-0e7679130b71-ca-trust-extracted\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.578448 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.578509 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-tls\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.578564 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c41ceaa-bc0a-4576-b309-0e7679130b71-installation-pull-secrets\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.578603 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-trusted-ca\") pod \"9c41ceaa-bc0a-4576-b309-0e7679130b71\" (UID: \"9c41ceaa-bc0a-4576-b309-0e7679130b71\") " Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.579219 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.579600 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.586963 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c41ceaa-bc0a-4576-b309-0e7679130b71-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.591434 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.591908 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-kube-api-access-tmm5k" (OuterVolumeSpecName: "kube-api-access-tmm5k") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "kube-api-access-tmm5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.592469 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.596366 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.608851 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c41ceaa-bc0a-4576-b309-0e7679130b71-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9c41ceaa-bc0a-4576-b309-0e7679130b71" (UID: "9c41ceaa-bc0a-4576-b309-0e7679130b71"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680772 4829 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680828 4829 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9c41ceaa-bc0a-4576-b309-0e7679130b71-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680851 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680869 4829 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9c41ceaa-bc0a-4576-b309-0e7679130b71-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680886 4829 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680905 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmm5k\" (UniqueName: \"kubernetes.io/projected/9c41ceaa-bc0a-4576-b309-0e7679130b71-kube-api-access-tmm5k\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:20 crc kubenswrapper[4829]: I1002 07:25:20.680924 4829 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9c41ceaa-bc0a-4576-b309-0e7679130b71-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.134050 4829 generic.go:334] "Generic (PLEG): container finished" podID="9c41ceaa-bc0a-4576-b309-0e7679130b71" containerID="b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c" exitCode=0 Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.134098 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" event={"ID":"9c41ceaa-bc0a-4576-b309-0e7679130b71","Type":"ContainerDied","Data":"b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c"} Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.134129 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" event={"ID":"9c41ceaa-bc0a-4576-b309-0e7679130b71","Type":"ContainerDied","Data":"2c85a608ab8876f259c8016732286b1ed45853ab3768844cb8290517ed832c61"} Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.134153 4829 scope.go:117] "RemoveContainer" containerID="b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c" Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.135281 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qx6xb" Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.163132 4829 scope.go:117] "RemoveContainer" containerID="b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c" Oct 02 07:25:21 crc kubenswrapper[4829]: E1002 07:25:21.163718 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c\": container with ID starting with b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c not found: ID does not exist" containerID="b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c" Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.163781 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c"} err="failed to get container status \"b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c\": rpc error: code = NotFound desc = could not find container \"b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c\": container with ID starting with b6a9fbb806ff299d602b59e2b778ebccdbfecd9893b153441690beeb73790c3c not found: ID does not exist" Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.196746 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qx6xb"] Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.204363 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qx6xb"] Oct 02 07:25:21 crc kubenswrapper[4829]: I1002 07:25:21.474564 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c41ceaa-bc0a-4576-b309-0e7679130b71" path="/var/lib/kubelet/pods/9c41ceaa-bc0a-4576-b309-0e7679130b71/volumes" Oct 02 07:25:49 crc kubenswrapper[4829]: I1002 07:25:49.669108 4829 scope.go:117] "RemoveContainer" containerID="ac66a22531034d0affa1529759b983ac2f8598f419b9bc97a55b6b616367e33c" Oct 02 07:25:55 crc kubenswrapper[4829]: I1002 07:25:55.329318 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:25:55 crc kubenswrapper[4829]: I1002 07:25:55.329724 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.448843 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-tc7zk"] Oct 02 07:26:12 crc kubenswrapper[4829]: E1002 07:26:12.449634 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c41ceaa-bc0a-4576-b309-0e7679130b71" containerName="registry" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.449651 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c41ceaa-bc0a-4576-b309-0e7679130b71" containerName="registry" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.449770 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c41ceaa-bc0a-4576-b309-0e7679130b71" containerName="registry" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.450340 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.452597 4829 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-m8mqw" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.453036 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.453626 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.460057 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxzhj\" (UniqueName: \"kubernetes.io/projected/13d4846f-c8e7-40a4-b312-3b3236539721-kube-api-access-mxzhj\") pod \"cert-manager-cainjector-7f985d654d-tc7zk\" (UID: \"13d4846f-c8e7-40a4-b312-3b3236539721\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.467699 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-tc7zk"] Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.470505 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d6q7t"] Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.471355 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d6q7t" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.473727 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-f6xpf"] Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.475792 4829 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-fttwk" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.475891 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.478200 4829 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-vj97t" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.489272 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d6q7t"] Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.492749 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-f6xpf"] Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.561550 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgzt\" (UniqueName: \"kubernetes.io/projected/adcd7359-41d5-433b-a4d1-aa964b09f997-kube-api-access-plgzt\") pod \"cert-manager-webhook-5655c58dd6-f6xpf\" (UID: \"adcd7359-41d5-433b-a4d1-aa964b09f997\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.561623 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhh2f\" (UniqueName: \"kubernetes.io/projected/482724e1-76a1-40b1-b5e2-1a51cade2778-kube-api-access-zhh2f\") pod \"cert-manager-5b446d88c5-d6q7t\" (UID: \"482724e1-76a1-40b1-b5e2-1a51cade2778\") " pod="cert-manager/cert-manager-5b446d88c5-d6q7t" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.561669 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxzhj\" (UniqueName: \"kubernetes.io/projected/13d4846f-c8e7-40a4-b312-3b3236539721-kube-api-access-mxzhj\") pod \"cert-manager-cainjector-7f985d654d-tc7zk\" (UID: \"13d4846f-c8e7-40a4-b312-3b3236539721\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.586302 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxzhj\" (UniqueName: \"kubernetes.io/projected/13d4846f-c8e7-40a4-b312-3b3236539721-kube-api-access-mxzhj\") pod \"cert-manager-cainjector-7f985d654d-tc7zk\" (UID: \"13d4846f-c8e7-40a4-b312-3b3236539721\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.662642 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgzt\" (UniqueName: \"kubernetes.io/projected/adcd7359-41d5-433b-a4d1-aa964b09f997-kube-api-access-plgzt\") pod \"cert-manager-webhook-5655c58dd6-f6xpf\" (UID: \"adcd7359-41d5-433b-a4d1-aa964b09f997\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.662691 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhh2f\" (UniqueName: \"kubernetes.io/projected/482724e1-76a1-40b1-b5e2-1a51cade2778-kube-api-access-zhh2f\") pod \"cert-manager-5b446d88c5-d6q7t\" (UID: \"482724e1-76a1-40b1-b5e2-1a51cade2778\") " pod="cert-manager/cert-manager-5b446d88c5-d6q7t" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.678800 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhh2f\" (UniqueName: \"kubernetes.io/projected/482724e1-76a1-40b1-b5e2-1a51cade2778-kube-api-access-zhh2f\") pod \"cert-manager-5b446d88c5-d6q7t\" (UID: \"482724e1-76a1-40b1-b5e2-1a51cade2778\") " pod="cert-manager/cert-manager-5b446d88c5-d6q7t" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.680976 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgzt\" (UniqueName: \"kubernetes.io/projected/adcd7359-41d5-433b-a4d1-aa964b09f997-kube-api-access-plgzt\") pod \"cert-manager-webhook-5655c58dd6-f6xpf\" (UID: \"adcd7359-41d5-433b-a4d1-aa964b09f997\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.768278 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.791587 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-d6q7t" Oct 02 07:26:12 crc kubenswrapper[4829]: I1002 07:26:12.797264 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.062587 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-d6q7t"] Oct 02 07:26:13 crc kubenswrapper[4829]: W1002 07:26:13.074884 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod482724e1_76a1_40b1_b5e2_1a51cade2778.slice/crio-ca3ab664f4900d85eac26713e691c1e9ee93dfbb7b96a3f70945013e720ebbbb WatchSource:0}: Error finding container ca3ab664f4900d85eac26713e691c1e9ee93dfbb7b96a3f70945013e720ebbbb: Status 404 returned error can't find the container with id ca3ab664f4900d85eac26713e691c1e9ee93dfbb7b96a3f70945013e720ebbbb Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.078344 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.227254 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-tc7zk"] Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.234299 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-f6xpf"] Oct 02 07:26:13 crc kubenswrapper[4829]: W1002 07:26:13.237199 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadcd7359_41d5_433b_a4d1_aa964b09f997.slice/crio-9628768a4bda6429eb97db15a3da6c7778aee4bae4aa7066b88ae130f212e155 WatchSource:0}: Error finding container 9628768a4bda6429eb97db15a3da6c7778aee4bae4aa7066b88ae130f212e155: Status 404 returned error can't find the container with id 9628768a4bda6429eb97db15a3da6c7778aee4bae4aa7066b88ae130f212e155 Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.470414 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" event={"ID":"adcd7359-41d5-433b-a4d1-aa964b09f997","Type":"ContainerStarted","Data":"9628768a4bda6429eb97db15a3da6c7778aee4bae4aa7066b88ae130f212e155"} Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.471645 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" event={"ID":"13d4846f-c8e7-40a4-b312-3b3236539721","Type":"ContainerStarted","Data":"05b88be22cd4c6450f700675b8366ba17d5bcc99dbb7fb8fa6c7385cdbe71ce2"} Oct 02 07:26:13 crc kubenswrapper[4829]: I1002 07:26:13.472736 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d6q7t" event={"ID":"482724e1-76a1-40b1-b5e2-1a51cade2778","Type":"ContainerStarted","Data":"ca3ab664f4900d85eac26713e691c1e9ee93dfbb7b96a3f70945013e720ebbbb"} Oct 02 07:26:15 crc kubenswrapper[4829]: I1002 07:26:15.483822 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-d6q7t" event={"ID":"482724e1-76a1-40b1-b5e2-1a51cade2778","Type":"ContainerStarted","Data":"9d343e253c79569aa58a297131ec0d53e9ed46a8ca3183b564017c3c7d2fb684"} Oct 02 07:26:15 crc kubenswrapper[4829]: I1002 07:26:15.504132 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-d6q7t" podStartSLOduration=1.413319729 podStartE2EDuration="3.504111975s" podCreationTimestamp="2025-10-02 07:26:12 +0000 UTC" firstStartedPulling="2025-10-02 07:26:13.078144723 +0000 UTC m=+564.417793118" lastFinishedPulling="2025-10-02 07:26:15.168936949 +0000 UTC m=+566.508585364" observedRunningTime="2025-10-02 07:26:15.502593106 +0000 UTC m=+566.842241521" watchObservedRunningTime="2025-10-02 07:26:15.504111975 +0000 UTC m=+566.843760390" Oct 02 07:26:17 crc kubenswrapper[4829]: I1002 07:26:17.501983 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" event={"ID":"13d4846f-c8e7-40a4-b312-3b3236539721","Type":"ContainerStarted","Data":"ad131fe703794c373951f9cf0ded5edec8e0c9b1eae4656e8fcf74abc12957ce"} Oct 02 07:26:17 crc kubenswrapper[4829]: I1002 07:26:17.503998 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" event={"ID":"adcd7359-41d5-433b-a4d1-aa964b09f997","Type":"ContainerStarted","Data":"bcb4506d7f09aebdf40d3e0fa94139ce1716f352cb54d9ca3aa4b08d88d9b238"} Oct 02 07:26:17 crc kubenswrapper[4829]: I1002 07:26:17.504483 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:17 crc kubenswrapper[4829]: I1002 07:26:17.530781 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-tc7zk" podStartSLOduration=2.246039342 podStartE2EDuration="5.53075711s" podCreationTimestamp="2025-10-02 07:26:12 +0000 UTC" firstStartedPulling="2025-10-02 07:26:13.23071044 +0000 UTC m=+564.570358855" lastFinishedPulling="2025-10-02 07:26:16.515428188 +0000 UTC m=+567.855076623" observedRunningTime="2025-10-02 07:26:17.525270072 +0000 UTC m=+568.864918497" watchObservedRunningTime="2025-10-02 07:26:17.53075711 +0000 UTC m=+568.870405525" Oct 02 07:26:17 crc kubenswrapper[4829]: I1002 07:26:17.550465 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" podStartSLOduration=2.265516702 podStartE2EDuration="5.550448276s" podCreationTimestamp="2025-10-02 07:26:12 +0000 UTC" firstStartedPulling="2025-10-02 07:26:13.240757975 +0000 UTC m=+564.580406390" lastFinishedPulling="2025-10-02 07:26:16.525689559 +0000 UTC m=+567.865337964" observedRunningTime="2025-10-02 07:26:17.549372121 +0000 UTC m=+568.889020576" watchObservedRunningTime="2025-10-02 07:26:17.550448276 +0000 UTC m=+568.890096681" Oct 02 07:26:22 crc kubenswrapper[4829]: I1002 07:26:22.801210 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-f6xpf" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.233676 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2lvjk"] Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.235029 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="nbdb" containerID="cri-o://8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.235467 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="northd" containerID="cri-o://580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.235721 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="sbdb" containerID="cri-o://52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.235025 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-controller" containerID="cri-o://39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.235884 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-acl-logging" containerID="cri-o://cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.235951 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-node" containerID="cri-o://412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.236329 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.297601 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" containerID="cri-o://714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d" gracePeriod=30 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.543701 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/2.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.544481 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/1.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.544524 4829 generic.go:334] "Generic (PLEG): container finished" podID="ca00dadf-1664-466b-830c-e172857db47a" containerID="5521c63b7a6584e5b77145dac98a81205b38da465e7c934ddd3fe3ad92d997be" exitCode=2 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.544577 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerDied","Data":"5521c63b7a6584e5b77145dac98a81205b38da465e7c934ddd3fe3ad92d997be"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.544615 4829 scope.go:117] "RemoveContainer" containerID="3d9d047b03925369a749b1ae75074f8d59e0a8c3d2c48252b537eec26461608c" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.545142 4829 scope.go:117] "RemoveContainer" containerID="5521c63b7a6584e5b77145dac98a81205b38da465e7c934ddd3fe3ad92d997be" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.545345 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-spzjt_openshift-multus(ca00dadf-1664-466b-830c-e172857db47a)\"" pod="openshift-multus/multus-spzjt" podUID="ca00dadf-1664-466b-830c-e172857db47a" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.548321 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/3.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.550699 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovn-acl-logging/0.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551143 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovn-controller/0.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551504 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d" exitCode=0 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551524 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95" exitCode=0 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551534 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c" exitCode=0 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551542 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d" exitCode=0 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551549 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844" exitCode=143 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551558 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0" exitCode=143 Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551576 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551600 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551610 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551620 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551630 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.551639 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0"} Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.579734 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovnkube-controller/3.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.582136 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovn-acl-logging/0.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.582673 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovn-controller/0.log" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.583150 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.585299 4829 scope.go:117] "RemoveContainer" containerID="f0cac1dd58f7920f79713074e1682848faaafde708e3ba97bfd94a72dbc37ff9" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.642819 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zkpdd"] Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643062 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643075 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643087 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-node" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643095 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-node" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643106 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="nbdb" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643111 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="nbdb" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643120 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643126 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643134 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="sbdb" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643140 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="sbdb" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643153 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kubecfg-setup" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643161 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kubecfg-setup" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643171 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-acl-logging" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643180 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-acl-logging" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643193 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643200 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643211 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="northd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643219 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="northd" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643246 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643253 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643261 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643267 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643275 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643281 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643372 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-node" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643382 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="northd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643390 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643396 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="sbdb" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643406 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643414 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643421 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643431 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-acl-logging" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643439 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643446 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovn-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643455 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="nbdb" Oct 02 07:26:23 crc kubenswrapper[4829]: E1002 07:26:23.643571 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643580 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.643664 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerName="ovnkube-controller" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.645204 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711485 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-kubelet\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711538 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-config\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711561 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-systemd-units\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711615 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-script-lib\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711650 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-openvswitch\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711666 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-node-log\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711657 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711684 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-etc-openvswitch\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711655 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711730 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711758 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-node-log" (OuterVolumeSpecName: "node-log") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711789 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711707 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4smrk\" (UniqueName: \"kubernetes.io/projected/00c0190d-cefe-408f-88c0-fa849e3811d3-kube-api-access-4smrk\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711854 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-slash\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711899 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-bin\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711963 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-slash" (OuterVolumeSpecName: "host-slash") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.711983 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712000 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00c0190d-cefe-408f-88c0-fa849e3811d3-ovn-node-metrics-cert\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712019 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-var-lib-openvswitch\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712035 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-netd\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712072 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712095 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-systemd\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712117 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-log-socket\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712133 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712159 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-log-socket" (OuterVolumeSpecName: "log-socket") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712180 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712200 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-ovn-kubernetes\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712259 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712217 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712294 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-netns\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712349 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712364 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712382 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-env-overrides\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712402 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-ovn\") pod \"00c0190d-cefe-408f-88c0-fa849e3811d3\" (UID: \"00c0190d-cefe-408f-88c0-fa849e3811d3\") " Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712510 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.712628 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713044 4829 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713058 4829 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713066 4829 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713074 4829 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713081 4829 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713089 4829 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713096 4829 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713103 4829 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713111 4829 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713119 4829 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713126 4829 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713135 4829 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713142 4829 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713151 4829 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713159 4829 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713167 4829 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.713210 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.716813 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00c0190d-cefe-408f-88c0-fa849e3811d3-kube-api-access-4smrk" (OuterVolumeSpecName: "kube-api-access-4smrk") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "kube-api-access-4smrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.717102 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00c0190d-cefe-408f-88c0-fa849e3811d3-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.732267 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "00c0190d-cefe-408f-88c0-fa849e3811d3" (UID: "00c0190d-cefe-408f-88c0-fa849e3811d3"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.814868 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-etc-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.814999 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-cni-netd\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815049 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815134 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-node-log\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815187 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-systemd-units\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815222 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-systemd\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815306 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815432 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-var-lib-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815531 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-log-socket\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815593 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovnkube-config\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815641 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovn-node-metrics-cert\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815727 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-run-ovn-kubernetes\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815789 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-slash\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.815967 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovnkube-script-lib\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816019 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-cni-bin\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816084 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-ovn\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816151 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-kubelet\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816210 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml5c2\" (UniqueName: \"kubernetes.io/projected/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-kube-api-access-ml5c2\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816279 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-run-netns\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816310 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-env-overrides\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816434 4829 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00c0190d-cefe-408f-88c0-fa849e3811d3-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816465 4829 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/00c0190d-cefe-408f-88c0-fa849e3811d3-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816489 4829 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/00c0190d-cefe-408f-88c0-fa849e3811d3-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.816510 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4smrk\" (UniqueName: \"kubernetes.io/projected/00c0190d-cefe-408f-88c0-fa849e3811d3-kube-api-access-4smrk\") on node \"crc\" DevicePath \"\"" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917587 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovnkube-script-lib\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917660 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-cni-bin\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917698 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-ovn\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917735 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-kubelet\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917785 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-cni-bin\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917816 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-ovn\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917887 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-kubelet\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.917942 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml5c2\" (UniqueName: \"kubernetes.io/projected/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-kube-api-access-ml5c2\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918050 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-run-netns\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918101 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-env-overrides\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918152 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-etc-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918190 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-cni-netd\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918262 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918315 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-node-log\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918371 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-systemd-units\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918407 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-systemd\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918456 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-var-lib-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918497 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-log-socket\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918538 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918584 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovnkube-config\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918625 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovn-node-metrics-cert\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918668 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-run-ovn-kubernetes\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918709 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-slash\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918798 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-env-overrides\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918835 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-run-netns\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918834 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-slash\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918858 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-systemd\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918878 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-etc-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918900 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-cni-netd\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918902 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-var-lib-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918920 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-run-openvswitch\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918939 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-node-log\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918958 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-systemd-units\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.918969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-log-socket\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.919251 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-run-ovn-kubernetes\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.919346 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.919413 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovnkube-config\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.919945 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovnkube-script-lib\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.923687 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-ovn-node-metrics-cert\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.945909 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml5c2\" (UniqueName: \"kubernetes.io/projected/b8516809-b4f3-47e4-8ead-d2e9eaf7360a-kube-api-access-ml5c2\") pod \"ovnkube-node-zkpdd\" (UID: \"b8516809-b4f3-47e4-8ead-d2e9eaf7360a\") " pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:23 crc kubenswrapper[4829]: I1002 07:26:23.964679 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.560508 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/2.log" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.566308 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovn-acl-logging/0.log" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.567203 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-2lvjk_00c0190d-cefe-408f-88c0-fa849e3811d3/ovn-controller/0.log" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568014 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a" exitCode=0 Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568055 4829 generic.go:334] "Generic (PLEG): container finished" podID="00c0190d-cefe-408f-88c0-fa849e3811d3" containerID="580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b" exitCode=0 Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568155 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a"} Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568262 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568310 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b"} Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568350 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-2lvjk" event={"ID":"00c0190d-cefe-408f-88c0-fa849e3811d3","Type":"ContainerDied","Data":"0ec09b6c64e5ce70c10c47472e451e836c791d55f25622d3490d257010b85834"} Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.568387 4829 scope.go:117] "RemoveContainer" containerID="714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.570350 4829 generic.go:334] "Generic (PLEG): container finished" podID="b8516809-b4f3-47e4-8ead-d2e9eaf7360a" containerID="06aa4ea04fc5a15a47cecc6d707dfc7807436ed2461d957d3d4e2dd4d70c4dd3" exitCode=0 Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.570388 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerDied","Data":"06aa4ea04fc5a15a47cecc6d707dfc7807436ed2461d957d3d4e2dd4d70c4dd3"} Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.570421 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"d34f187c2a8dbb75e93c024c4e435aa7a767baea004dc3a50ffc0f305783b3c2"} Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.604920 4829 scope.go:117] "RemoveContainer" containerID="52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.621567 4829 scope.go:117] "RemoveContainer" containerID="8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.654537 4829 scope.go:117] "RemoveContainer" containerID="580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.681995 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2lvjk"] Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.696698 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-2lvjk"] Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.699751 4829 scope.go:117] "RemoveContainer" containerID="3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.738366 4829 scope.go:117] "RemoveContainer" containerID="412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.753754 4829 scope.go:117] "RemoveContainer" containerID="cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.767217 4829 scope.go:117] "RemoveContainer" containerID="39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.782693 4829 scope.go:117] "RemoveContainer" containerID="a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.804515 4829 scope.go:117] "RemoveContainer" containerID="714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.805183 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d\": container with ID starting with 714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d not found: ID does not exist" containerID="714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.805250 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d"} err="failed to get container status \"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d\": rpc error: code = NotFound desc = could not find container \"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d\": container with ID starting with 714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.805277 4829 scope.go:117] "RemoveContainer" containerID="52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.805816 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\": container with ID starting with 52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a not found: ID does not exist" containerID="52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.805885 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a"} err="failed to get container status \"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\": rpc error: code = NotFound desc = could not find container \"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\": container with ID starting with 52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.805930 4829 scope.go:117] "RemoveContainer" containerID="8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.806460 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\": container with ID starting with 8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95 not found: ID does not exist" containerID="8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.806520 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95"} err="failed to get container status \"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\": rpc error: code = NotFound desc = could not find container \"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\": container with ID starting with 8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.806542 4829 scope.go:117] "RemoveContainer" containerID="580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.806995 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\": container with ID starting with 580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b not found: ID does not exist" containerID="580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.807045 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b"} err="failed to get container status \"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\": rpc error: code = NotFound desc = could not find container \"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\": container with ID starting with 580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.807077 4829 scope.go:117] "RemoveContainer" containerID="3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.807573 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\": container with ID starting with 3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c not found: ID does not exist" containerID="3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.807730 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c"} err="failed to get container status \"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\": rpc error: code = NotFound desc = could not find container \"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\": container with ID starting with 3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.807772 4829 scope.go:117] "RemoveContainer" containerID="412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.808121 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\": container with ID starting with 412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d not found: ID does not exist" containerID="412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.808153 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d"} err="failed to get container status \"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\": rpc error: code = NotFound desc = could not find container \"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\": container with ID starting with 412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.808173 4829 scope.go:117] "RemoveContainer" containerID="cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.808541 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\": container with ID starting with cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844 not found: ID does not exist" containerID="cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.808599 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844"} err="failed to get container status \"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\": rpc error: code = NotFound desc = could not find container \"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\": container with ID starting with cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.808621 4829 scope.go:117] "RemoveContainer" containerID="39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.809105 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\": container with ID starting with 39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0 not found: ID does not exist" containerID="39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.809139 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0"} err="failed to get container status \"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\": rpc error: code = NotFound desc = could not find container \"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\": container with ID starting with 39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.809159 4829 scope.go:117] "RemoveContainer" containerID="a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800" Oct 02 07:26:24 crc kubenswrapper[4829]: E1002 07:26:24.809551 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\": container with ID starting with a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800 not found: ID does not exist" containerID="a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.809584 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800"} err="failed to get container status \"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\": rpc error: code = NotFound desc = could not find container \"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\": container with ID starting with a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.809603 4829 scope.go:117] "RemoveContainer" containerID="714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.809881 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d"} err="failed to get container status \"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d\": rpc error: code = NotFound desc = could not find container \"714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d\": container with ID starting with 714142942e7c033ffad89e0716fd8567edac753f2660a1cac15cac7419e1ba9d not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.809906 4829 scope.go:117] "RemoveContainer" containerID="52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.810494 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a"} err="failed to get container status \"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\": rpc error: code = NotFound desc = could not find container \"52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a\": container with ID starting with 52b373b7d142f5c8f1d7700a7cbaf92add30e8d2a8fb290525c7bd5ff3d0d02a not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.810525 4829 scope.go:117] "RemoveContainer" containerID="8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.810880 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95"} err="failed to get container status \"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\": rpc error: code = NotFound desc = could not find container \"8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95\": container with ID starting with 8198df61440042c7ced4a4b082a84765f34fd9187dde3d367c34d7290a920f95 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.810908 4829 scope.go:117] "RemoveContainer" containerID="580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.811202 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b"} err="failed to get container status \"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\": rpc error: code = NotFound desc = could not find container \"580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b\": container with ID starting with 580ff0df1b394653f807a249935020062f73b0c3c61d76d252af00d2d0d3979b not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.811429 4829 scope.go:117] "RemoveContainer" containerID="3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.811983 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c"} err="failed to get container status \"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\": rpc error: code = NotFound desc = could not find container \"3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c\": container with ID starting with 3741c36308e4f067e503a0215cd12f21269ee97e40bc6e1073bd6a1fdb10aa8c not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.812016 4829 scope.go:117] "RemoveContainer" containerID="412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.812321 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d"} err="failed to get container status \"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\": rpc error: code = NotFound desc = could not find container \"412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d\": container with ID starting with 412bb0fd9c76c86eab2648d493b896a536dcff7aa18166703971f46353cce53d not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.812345 4829 scope.go:117] "RemoveContainer" containerID="cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.812637 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844"} err="failed to get container status \"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\": rpc error: code = NotFound desc = could not find container \"cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844\": container with ID starting with cba03c46528dfc6f6ab4d5b021a9ce6836a70303f4b03621e8d2a6080fd75844 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.812679 4829 scope.go:117] "RemoveContainer" containerID="39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.813034 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0"} err="failed to get container status \"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\": rpc error: code = NotFound desc = could not find container \"39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0\": container with ID starting with 39728fad5cfd142138ed73d78419eddf8967d88f7bf13fba9dcc9dbda33fbef0 not found: ID does not exist" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.813062 4829 scope.go:117] "RemoveContainer" containerID="a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800" Oct 02 07:26:24 crc kubenswrapper[4829]: I1002 07:26:24.813335 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800"} err="failed to get container status \"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\": rpc error: code = NotFound desc = could not find container \"a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800\": container with ID starting with a1f78124513531e8c6c1f8498a37e3935ecd122833c621b4ba1c80cd555b2800 not found: ID does not exist" Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.328660 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.329026 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.470191 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00c0190d-cefe-408f-88c0-fa849e3811d3" path="/var/lib/kubelet/pods/00c0190d-cefe-408f-88c0-fa849e3811d3/volumes" Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.586555 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"3627d84d6a74598f00946ab7a994dec24127fe720b3e6b9e4e1817eebf911b71"} Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.586608 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"e80a6c7d91fb1649f8e8ef99879de007596525be9c214b62c456b6fe7db99713"} Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.586632 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"850552e3d0d76fc8b4204f54594aec41d6036b0e65b7c17f43e2afdfe2d0179f"} Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.586651 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"1c14cd324922deaea1da64378958db3a5ca2591731c3fe791670ec9999823a6a"} Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.586668 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"d71be09c5f091cf4463c9b69e65202da17e981ec613564d76622e62e2df04b05"} Oct 02 07:26:25 crc kubenswrapper[4829]: I1002 07:26:25.586685 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"e88d9987a5ccee57b9ef909c2f47a13ea31495323796cc7e58b64c9cea63c8f2"} Oct 02 07:26:28 crc kubenswrapper[4829]: I1002 07:26:28.615035 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"10907478c198129b4be86211fb05f36f1d406c9a4308949e10338dfb85238fc4"} Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.633948 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" event={"ID":"b8516809-b4f3-47e4-8ead-d2e9eaf7360a","Type":"ContainerStarted","Data":"ed28148d78e46a714807c7c12366c0f22c7d026851e0c236711732aab90a8c17"} Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.635736 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.635767 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.635777 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.660022 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.666075 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:30 crc kubenswrapper[4829]: I1002 07:26:30.670348 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" podStartSLOduration=7.670337946 podStartE2EDuration="7.670337946s" podCreationTimestamp="2025-10-02 07:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:26:30.668263239 +0000 UTC m=+582.007911654" watchObservedRunningTime="2025-10-02 07:26:30.670337946 +0000 UTC m=+582.009986351" Oct 02 07:26:36 crc kubenswrapper[4829]: I1002 07:26:36.460728 4829 scope.go:117] "RemoveContainer" containerID="5521c63b7a6584e5b77145dac98a81205b38da465e7c934ddd3fe3ad92d997be" Oct 02 07:26:36 crc kubenswrapper[4829]: E1002 07:26:36.461818 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-spzjt_openshift-multus(ca00dadf-1664-466b-830c-e172857db47a)\"" pod="openshift-multus/multus-spzjt" podUID="ca00dadf-1664-466b-830c-e172857db47a" Oct 02 07:26:50 crc kubenswrapper[4829]: I1002 07:26:50.974481 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w"] Oct 02 07:26:50 crc kubenswrapper[4829]: I1002 07:26:50.976485 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:50 crc kubenswrapper[4829]: I1002 07:26:50.978876 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 07:26:50 crc kubenswrapper[4829]: I1002 07:26:50.987859 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w"] Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.132417 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.132590 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.132763 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdzzt\" (UniqueName: \"kubernetes.io/projected/ff58a0e1-dfff-438b-bd40-00240f7372dd-kube-api-access-hdzzt\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.233681 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdzzt\" (UniqueName: \"kubernetes.io/projected/ff58a0e1-dfff-438b-bd40-00240f7372dd-kube-api-access-hdzzt\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.233754 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.233787 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.234196 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.234453 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.251953 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdzzt\" (UniqueName: \"kubernetes.io/projected/ff58a0e1-dfff-438b-bd40-00240f7372dd-kube-api-access-hdzzt\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.307547 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.349802 4829 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(0f71a176d355b3b83ba515632e4574fe8050372ba64f625bf15a087282f09351): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.349910 4829 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(0f71a176d355b3b83ba515632e4574fe8050372ba64f625bf15a087282f09351): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.349948 4829 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(0f71a176d355b3b83ba515632e4574fe8050372ba64f625bf15a087282f09351): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.350049 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace(ff58a0e1-dfff-438b-bd40-00240f7372dd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace(ff58a0e1-dfff-438b-bd40-00240f7372dd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(0f71a176d355b3b83ba515632e4574fe8050372ba64f625bf15a087282f09351): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.460897 4829 scope.go:117] "RemoveContainer" containerID="5521c63b7a6584e5b77145dac98a81205b38da465e7c934ddd3fe3ad92d997be" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.784394 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-spzjt_ca00dadf-1664-466b-830c-e172857db47a/kube-multus/2.log" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.784981 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-spzjt" event={"ID":"ca00dadf-1664-466b-830c-e172857db47a","Type":"ContainerStarted","Data":"c078da6cf2359e8ddd2c0d14a25ee30a02f37fc0d61b6bde40e4e3763b3be31a"} Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.785199 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: I1002 07:26:51.786143 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.852305 4829 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(c0dc54f2d46db570b65b4ef88dc53b3b0e2a2db2f9fd16783603db760a4151c3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.852407 4829 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(c0dc54f2d46db570b65b4ef88dc53b3b0e2a2db2f9fd16783603db760a4151c3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.852446 4829 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(c0dc54f2d46db570b65b4ef88dc53b3b0e2a2db2f9fd16783603db760a4151c3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:26:51 crc kubenswrapper[4829]: E1002 07:26:51.852527 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace(ff58a0e1-dfff-438b-bd40-00240f7372dd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace(ff58a0e1-dfff-438b-bd40-00240f7372dd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_openshift-marketplace_ff58a0e1-dfff-438b-bd40-00240f7372dd_0(c0dc54f2d46db570b65b4ef88dc53b3b0e2a2db2f9fd16783603db760a4151c3): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" Oct 02 07:26:54 crc kubenswrapper[4829]: I1002 07:26:53.998761 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zkpdd" Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.329114 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.329818 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.329944 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.330615 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c344e588de3b2b45a68b8eaf305327aebb97a91305e92834159bf1c925efb2c"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.330761 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://3c344e588de3b2b45a68b8eaf305327aebb97a91305e92834159bf1c925efb2c" gracePeriod=600 Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.811171 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="3c344e588de3b2b45a68b8eaf305327aebb97a91305e92834159bf1c925efb2c" exitCode=0 Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.811274 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"3c344e588de3b2b45a68b8eaf305327aebb97a91305e92834159bf1c925efb2c"} Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.811545 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"d8e666d4540458f8abd1af2cd0bebf671777110d93b151c9aedce3fa191263d9"} Oct 02 07:26:55 crc kubenswrapper[4829]: I1002 07:26:55.811572 4829 scope.go:117] "RemoveContainer" containerID="eaf4b9893a38009f8a714f08d03494fe5d6ecc601b0a3a024d68f3487caee9a2" Oct 02 07:27:02 crc kubenswrapper[4829]: I1002 07:27:02.460516 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:27:02 crc kubenswrapper[4829]: I1002 07:27:02.461877 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:27:02 crc kubenswrapper[4829]: I1002 07:27:02.738833 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w"] Oct 02 07:27:02 crc kubenswrapper[4829]: I1002 07:27:02.869335 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" event={"ID":"ff58a0e1-dfff-438b-bd40-00240f7372dd","Type":"ContainerStarted","Data":"a3bece0835b810c8e229f5e4a63c51075a4102b53c82a61a0f11b48ba473f9e0"} Oct 02 07:27:03 crc kubenswrapper[4829]: I1002 07:27:03.879674 4829 generic.go:334] "Generic (PLEG): container finished" podID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerID="76b8f858fafa5ca6a9a4bc4c28b1a2baedcaf31a254f0c0f21476dd300c5ab35" exitCode=0 Oct 02 07:27:03 crc kubenswrapper[4829]: I1002 07:27:03.879743 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" event={"ID":"ff58a0e1-dfff-438b-bd40-00240f7372dd","Type":"ContainerDied","Data":"76b8f858fafa5ca6a9a4bc4c28b1a2baedcaf31a254f0c0f21476dd300c5ab35"} Oct 02 07:27:05 crc kubenswrapper[4829]: I1002 07:27:05.896690 4829 generic.go:334] "Generic (PLEG): container finished" podID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerID="863ccaf7e7022c66c8e464f84e224be2804b2ff3c7f9815b954d0b47a4977fa0" exitCode=0 Oct 02 07:27:05 crc kubenswrapper[4829]: I1002 07:27:05.896857 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" event={"ID":"ff58a0e1-dfff-438b-bd40-00240f7372dd","Type":"ContainerDied","Data":"863ccaf7e7022c66c8e464f84e224be2804b2ff3c7f9815b954d0b47a4977fa0"} Oct 02 07:27:06 crc kubenswrapper[4829]: I1002 07:27:06.910956 4829 generic.go:334] "Generic (PLEG): container finished" podID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerID="e5086192b466e260a433d3d2ea0b77d98d4a7b0af228ec0e1bb5ba2d5cd23602" exitCode=0 Oct 02 07:27:06 crc kubenswrapper[4829]: I1002 07:27:06.911018 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" event={"ID":"ff58a0e1-dfff-438b-bd40-00240f7372dd","Type":"ContainerDied","Data":"e5086192b466e260a433d3d2ea0b77d98d4a7b0af228ec0e1bb5ba2d5cd23602"} Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.163990 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.272611 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-util\") pod \"ff58a0e1-dfff-438b-bd40-00240f7372dd\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.272673 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-bundle\") pod \"ff58a0e1-dfff-438b-bd40-00240f7372dd\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.272701 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdzzt\" (UniqueName: \"kubernetes.io/projected/ff58a0e1-dfff-438b-bd40-00240f7372dd-kube-api-access-hdzzt\") pod \"ff58a0e1-dfff-438b-bd40-00240f7372dd\" (UID: \"ff58a0e1-dfff-438b-bd40-00240f7372dd\") " Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.276700 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-bundle" (OuterVolumeSpecName: "bundle") pod "ff58a0e1-dfff-438b-bd40-00240f7372dd" (UID: "ff58a0e1-dfff-438b-bd40-00240f7372dd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.278020 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff58a0e1-dfff-438b-bd40-00240f7372dd-kube-api-access-hdzzt" (OuterVolumeSpecName: "kube-api-access-hdzzt") pod "ff58a0e1-dfff-438b-bd40-00240f7372dd" (UID: "ff58a0e1-dfff-438b-bd40-00240f7372dd"). InnerVolumeSpecName "kube-api-access-hdzzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.374147 4829 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.374192 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdzzt\" (UniqueName: \"kubernetes.io/projected/ff58a0e1-dfff-438b-bd40-00240f7372dd-kube-api-access-hdzzt\") on node \"crc\" DevicePath \"\"" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.453835 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-util" (OuterVolumeSpecName: "util") pod "ff58a0e1-dfff-438b-bd40-00240f7372dd" (UID: "ff58a0e1-dfff-438b-bd40-00240f7372dd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.475648 4829 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ff58a0e1-dfff-438b-bd40-00240f7372dd-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.929507 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" event={"ID":"ff58a0e1-dfff-438b-bd40-00240f7372dd","Type":"ContainerDied","Data":"a3bece0835b810c8e229f5e4a63c51075a4102b53c82a61a0f11b48ba473f9e0"} Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.929560 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3bece0835b810c8e229f5e4a63c51075a4102b53c82a61a0f11b48ba473f9e0" Oct 02 07:27:08 crc kubenswrapper[4829]: I1002 07:27:08.929646 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.267691 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-448hn"] Oct 02 07:27:19 crc kubenswrapper[4829]: E1002 07:27:19.268247 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="extract" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.268258 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="extract" Oct 02 07:27:19 crc kubenswrapper[4829]: E1002 07:27:19.268269 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="util" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.268275 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="util" Oct 02 07:27:19 crc kubenswrapper[4829]: E1002 07:27:19.268288 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="pull" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.268293 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="pull" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.268381 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff58a0e1-dfff-438b-bd40-00240f7372dd" containerName="extract" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.268698 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" Oct 02 07:27:19 crc kubenswrapper[4829]: W1002 07:27:19.270162 4829 reflector.go:561] object-"openshift-operators"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Oct 02 07:27:19 crc kubenswrapper[4829]: E1002 07:27:19.270201 4829 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:27:19 crc kubenswrapper[4829]: W1002 07:27:19.270406 4829 reflector.go:561] object-"openshift-operators"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-operators": no relationship found between node 'crc' and this object Oct 02 07:27:19 crc kubenswrapper[4829]: E1002 07:27:19.270457 4829 reflector.go:158] "Unhandled Error" err="object-\"openshift-operators\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.271350 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-2x6p2" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.281897 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-448hn"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.385238 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.386050 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.388831 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.388838 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-xhfct" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.400179 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.401028 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.429868 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.434187 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpq6x\" (UniqueName: \"kubernetes.io/projected/873d1771-f008-4ac1-8ece-e3b48ba37a81-kube-api-access-qpq6x\") pod \"obo-prometheus-operator-7c8cf85677-448hn\" (UID: \"873d1771-f008-4ac1-8ece-e3b48ba37a81\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.434255 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f5f2c2d-e40c-4b14-97df-a5a83489726a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv\" (UID: \"9f5f2c2d-e40c-4b14-97df-a5a83489726a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.434290 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f5f2c2d-e40c-4b14-97df-a5a83489726a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv\" (UID: \"9f5f2c2d-e40c-4b14-97df-a5a83489726a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.457551 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.535335 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpq6x\" (UniqueName: \"kubernetes.io/projected/873d1771-f008-4ac1-8ece-e3b48ba37a81-kube-api-access-qpq6x\") pod \"obo-prometheus-operator-7c8cf85677-448hn\" (UID: \"873d1771-f008-4ac1-8ece-e3b48ba37a81\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.535712 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f5f2c2d-e40c-4b14-97df-a5a83489726a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv\" (UID: \"9f5f2c2d-e40c-4b14-97df-a5a83489726a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.535757 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1ddcd62-beba-4339-9aa1-df07ce7577a8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz\" (UID: \"c1ddcd62-beba-4339-9aa1-df07ce7577a8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.535794 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f5f2c2d-e40c-4b14-97df-a5a83489726a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv\" (UID: \"9f5f2c2d-e40c-4b14-97df-a5a83489726a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.535847 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1ddcd62-beba-4339-9aa1-df07ce7577a8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz\" (UID: \"c1ddcd62-beba-4339-9aa1-df07ce7577a8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.542707 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9f5f2c2d-e40c-4b14-97df-a5a83489726a-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv\" (UID: \"9f5f2c2d-e40c-4b14-97df-a5a83489726a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.549193 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9f5f2c2d-e40c-4b14-97df-a5a83489726a-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv\" (UID: \"9f5f2c2d-e40c-4b14-97df-a5a83489726a\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.597665 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-6th2l"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.598388 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.599827 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-mtnf6" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.600031 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.612462 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-6th2l"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.636913 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1ddcd62-beba-4339-9aa1-df07ce7577a8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz\" (UID: \"c1ddcd62-beba-4339-9aa1-df07ce7577a8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.636981 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1ddcd62-beba-4339-9aa1-df07ce7577a8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz\" (UID: \"c1ddcd62-beba-4339-9aa1-df07ce7577a8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.637005 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9175ba03-d3bd-4278-8943-1499754f510b-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-6th2l\" (UID: \"9175ba03-d3bd-4278-8943-1499754f510b\") " pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.637053 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/9175ba03-d3bd-4278-8943-1499754f510b-kube-api-access-q7phq\") pod \"observability-operator-cc5f78dfc-6th2l\" (UID: \"9175ba03-d3bd-4278-8943-1499754f510b\") " pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.639574 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c1ddcd62-beba-4339-9aa1-df07ce7577a8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz\" (UID: \"c1ddcd62-beba-4339-9aa1-df07ce7577a8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.639949 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c1ddcd62-beba-4339-9aa1-df07ce7577a8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz\" (UID: \"c1ddcd62-beba-4339-9aa1-df07ce7577a8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.702050 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.716439 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.737833 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9175ba03-d3bd-4278-8943-1499754f510b-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-6th2l\" (UID: \"9175ba03-d3bd-4278-8943-1499754f510b\") " pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.737903 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/9175ba03-d3bd-4278-8943-1499754f510b-kube-api-access-q7phq\") pod \"observability-operator-cc5f78dfc-6th2l\" (UID: \"9175ba03-d3bd-4278-8943-1499754f510b\") " pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.741371 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/9175ba03-d3bd-4278-8943-1499754f510b-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-6th2l\" (UID: \"9175ba03-d3bd-4278-8943-1499754f510b\") " pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.793147 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-mw7fx"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.794038 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.798332 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-4ljx2" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.811539 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-mw7fx"] Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.940839 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bb9v\" (UniqueName: \"kubernetes.io/projected/eb1be399-f44c-4acb-81ad-9723071d7f9f-kube-api-access-5bb9v\") pod \"perses-operator-54bc95c9fb-mw7fx\" (UID: \"eb1be399-f44c-4acb-81ad-9723071d7f9f\") " pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.941145 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1be399-f44c-4acb-81ad-9723071d7f9f-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-mw7fx\" (UID: \"eb1be399-f44c-4acb-81ad-9723071d7f9f\") " pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:19 crc kubenswrapper[4829]: I1002 07:27:19.969849 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv"] Oct 02 07:27:19 crc kubenswrapper[4829]: W1002 07:27:19.979103 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f5f2c2d_e40c_4b14_97df_a5a83489726a.slice/crio-ebf9fbaaa130a3a4d427a88abd3b68a4b4f7102246be8f7e62e3ec7876548f4c WatchSource:0}: Error finding container ebf9fbaaa130a3a4d427a88abd3b68a4b4f7102246be8f7e62e3ec7876548f4c: Status 404 returned error can't find the container with id ebf9fbaaa130a3a4d427a88abd3b68a4b4f7102246be8f7e62e3ec7876548f4c Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.018953 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" event={"ID":"9f5f2c2d-e40c-4b14-97df-a5a83489726a","Type":"ContainerStarted","Data":"ebf9fbaaa130a3a4d427a88abd3b68a4b4f7102246be8f7e62e3ec7876548f4c"} Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.032941 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz"] Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.044443 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1be399-f44c-4acb-81ad-9723071d7f9f-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-mw7fx\" (UID: \"eb1be399-f44c-4acb-81ad-9723071d7f9f\") " pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.044483 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bb9v\" (UniqueName: \"kubernetes.io/projected/eb1be399-f44c-4acb-81ad-9723071d7f9f-kube-api-access-5bb9v\") pod \"perses-operator-54bc95c9fb-mw7fx\" (UID: \"eb1be399-f44c-4acb-81ad-9723071d7f9f\") " pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.309430 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.316529 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1be399-f44c-4acb-81ad-9723071d7f9f-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-mw7fx\" (UID: \"eb1be399-f44c-4acb-81ad-9723071d7f9f\") " pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.465255 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.471948 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpq6x\" (UniqueName: \"kubernetes.io/projected/873d1771-f008-4ac1-8ece-e3b48ba37a81-kube-api-access-qpq6x\") pod \"obo-prometheus-operator-7c8cf85677-448hn\" (UID: \"873d1771-f008-4ac1-8ece-e3b48ba37a81\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.475218 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bb9v\" (UniqueName: \"kubernetes.io/projected/eb1be399-f44c-4acb-81ad-9723071d7f9f-kube-api-access-5bb9v\") pod \"perses-operator-54bc95c9fb-mw7fx\" (UID: \"eb1be399-f44c-4acb-81ad-9723071d7f9f\") " pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.476737 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7phq\" (UniqueName: \"kubernetes.io/projected/9175ba03-d3bd-4278-8943-1499754f510b-kube-api-access-q7phq\") pod \"observability-operator-cc5f78dfc-6th2l\" (UID: \"9175ba03-d3bd-4278-8943-1499754f510b\") " pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.486761 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.515316 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.716780 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.806478 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-448hn"] Oct 02 07:27:20 crc kubenswrapper[4829]: W1002 07:27:20.821650 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod873d1771_f008_4ac1_8ece_e3b48ba37a81.slice/crio-a1d40fc6071e064817dcabe93de06037f0c4f2dc2063084d06a28ed6e98ea93f WatchSource:0}: Error finding container a1d40fc6071e064817dcabe93de06037f0c4f2dc2063084d06a28ed6e98ea93f: Status 404 returned error can't find the container with id a1d40fc6071e064817dcabe93de06037f0c4f2dc2063084d06a28ed6e98ea93f Oct 02 07:27:20 crc kubenswrapper[4829]: I1002 07:27:20.861289 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-6th2l"] Oct 02 07:27:21 crc kubenswrapper[4829]: I1002 07:27:21.027504 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" event={"ID":"9175ba03-d3bd-4278-8943-1499754f510b","Type":"ContainerStarted","Data":"ae46f2d4b729ad388094396b47417e742dcbc27f4dfbf41eb7efc51be478eb50"} Oct 02 07:27:21 crc kubenswrapper[4829]: I1002 07:27:21.029831 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" event={"ID":"873d1771-f008-4ac1-8ece-e3b48ba37a81","Type":"ContainerStarted","Data":"a1d40fc6071e064817dcabe93de06037f0c4f2dc2063084d06a28ed6e98ea93f"} Oct 02 07:27:21 crc kubenswrapper[4829]: I1002 07:27:21.044053 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" event={"ID":"c1ddcd62-beba-4339-9aa1-df07ce7577a8","Type":"ContainerStarted","Data":"37dc6b4dc0f88c5927d8e0beaa26ff7f58c53e7551428bc7d64dce6208cd8110"} Oct 02 07:27:21 crc kubenswrapper[4829]: I1002 07:27:21.083839 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-mw7fx"] Oct 02 07:27:21 crc kubenswrapper[4829]: W1002 07:27:21.098488 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb1be399_f44c_4acb_81ad_9723071d7f9f.slice/crio-48695dd4f8df839ac12e99bc727c5ced4eb9b644ea0b7b96d0a537701b74f4c2 WatchSource:0}: Error finding container 48695dd4f8df839ac12e99bc727c5ced4eb9b644ea0b7b96d0a537701b74f4c2: Status 404 returned error can't find the container with id 48695dd4f8df839ac12e99bc727c5ced4eb9b644ea0b7b96d0a537701b74f4c2 Oct 02 07:27:22 crc kubenswrapper[4829]: I1002 07:27:22.068053 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" event={"ID":"eb1be399-f44c-4acb-81ad-9723071d7f9f","Type":"ContainerStarted","Data":"48695dd4f8df839ac12e99bc727c5ced4eb9b644ea0b7b96d0a537701b74f4c2"} Oct 02 07:27:34 crc kubenswrapper[4829]: E1002 07:27:34.179539 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c" Oct 02 07:27:34 crc kubenswrapper[4829]: E1002 07:27:34.180201 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5bb9v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-54bc95c9fb-mw7fx_openshift-operators(eb1be399-f44c-4acb-81ad-9723071d7f9f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 07:27:34 crc kubenswrapper[4829]: E1002 07:27:34.181534 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" podUID="eb1be399-f44c-4acb-81ad-9723071d7f9f" Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.183449 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" event={"ID":"873d1771-f008-4ac1-8ece-e3b48ba37a81","Type":"ContainerStarted","Data":"5a3efe18f3336f0c409edbf49a1558c3e5a975d3775bfb15dc6c363749dede57"} Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.185298 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" event={"ID":"c1ddcd62-beba-4339-9aa1-df07ce7577a8","Type":"ContainerStarted","Data":"0df049df726c7d5c001d6b51161c08e183ba72748a3f11db604f15bd128890c7"} Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.186813 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" event={"ID":"9f5f2c2d-e40c-4b14-97df-a5a83489726a","Type":"ContainerStarted","Data":"d3007391bbd97a101643963d514e4a36c1b54a356a18ed0a239457beadff995a"} Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.188918 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" event={"ID":"9175ba03-d3bd-4278-8943-1499754f510b","Type":"ContainerStarted","Data":"824951da348d81597462eb975cd22b13560d814d96056e1e7c5fd8cfe6b4b12e"} Oct 02 07:27:35 crc kubenswrapper[4829]: E1002 07:27:35.191156 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c\\\"\"" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" podUID="eb1be399-f44c-4acb-81ad-9723071d7f9f" Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.212581 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-448hn" podStartSLOduration=2.849027054 podStartE2EDuration="16.212561393s" podCreationTimestamp="2025-10-02 07:27:19 +0000 UTC" firstStartedPulling="2025-10-02 07:27:20.823671853 +0000 UTC m=+632.163320258" lastFinishedPulling="2025-10-02 07:27:34.187206192 +0000 UTC m=+645.526854597" observedRunningTime="2025-10-02 07:27:35.209217547 +0000 UTC m=+646.548865972" watchObservedRunningTime="2025-10-02 07:27:35.212561393 +0000 UTC m=+646.552209808" Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.253802 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz" podStartSLOduration=2.089313424 podStartE2EDuration="16.253784886s" podCreationTimestamp="2025-10-02 07:27:19 +0000 UTC" firstStartedPulling="2025-10-02 07:27:20.043124855 +0000 UTC m=+631.382773280" lastFinishedPulling="2025-10-02 07:27:34.207596337 +0000 UTC m=+645.547244742" observedRunningTime="2025-10-02 07:27:35.251512465 +0000 UTC m=+646.591160870" watchObservedRunningTime="2025-10-02 07:27:35.253784886 +0000 UTC m=+646.593433281" Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.278134 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" podStartSLOduration=2.975598596 podStartE2EDuration="16.278118256s" podCreationTimestamp="2025-10-02 07:27:19 +0000 UTC" firstStartedPulling="2025-10-02 07:27:20.903044943 +0000 UTC m=+632.242693348" lastFinishedPulling="2025-10-02 07:27:34.205564603 +0000 UTC m=+645.545213008" observedRunningTime="2025-10-02 07:27:35.27378791 +0000 UTC m=+646.613436315" watchObservedRunningTime="2025-10-02 07:27:35.278118256 +0000 UTC m=+646.617766661" Oct 02 07:27:35 crc kubenswrapper[4829]: I1002 07:27:35.327277 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv" podStartSLOduration=2.122898238 podStartE2EDuration="16.327259201s" podCreationTimestamp="2025-10-02 07:27:19 +0000 UTC" firstStartedPulling="2025-10-02 07:27:19.98983711 +0000 UTC m=+631.329485515" lastFinishedPulling="2025-10-02 07:27:34.194198053 +0000 UTC m=+645.533846478" observedRunningTime="2025-10-02 07:27:35.319402281 +0000 UTC m=+646.659050706" watchObservedRunningTime="2025-10-02 07:27:35.327259201 +0000 UTC m=+646.666907616" Oct 02 07:27:36 crc kubenswrapper[4829]: I1002 07:27:36.195998 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:36 crc kubenswrapper[4829]: I1002 07:27:36.200643 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-6th2l" Oct 02 07:27:51 crc kubenswrapper[4829]: I1002 07:27:51.299453 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" event={"ID":"eb1be399-f44c-4acb-81ad-9723071d7f9f","Type":"ContainerStarted","Data":"aca1977c0c0958218f5f077ff867a58da5ed3d39785c886623712218783ce15b"} Oct 02 07:27:51 crc kubenswrapper[4829]: I1002 07:27:51.300744 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:27:51 crc kubenswrapper[4829]: I1002 07:27:51.327471 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" podStartSLOduration=3.154760857 podStartE2EDuration="32.327442764s" podCreationTimestamp="2025-10-02 07:27:19 +0000 UTC" firstStartedPulling="2025-10-02 07:27:21.101189001 +0000 UTC m=+632.440837406" lastFinishedPulling="2025-10-02 07:27:50.273870898 +0000 UTC m=+661.613519313" observedRunningTime="2025-10-02 07:27:51.325533465 +0000 UTC m=+662.665181910" watchObservedRunningTime="2025-10-02 07:27:51.327442764 +0000 UTC m=+662.667091219" Oct 02 07:28:00 crc kubenswrapper[4829]: I1002 07:28:00.720901 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-mw7fx" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.124047 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm"] Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.126364 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.129040 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.145791 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm"] Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.297983 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.298128 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.298215 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rmb8\" (UniqueName: \"kubernetes.io/projected/d840464b-b486-4421-acd8-fded53601478-kube-api-access-5rmb8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.399903 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.399956 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rmb8\" (UniqueName: \"kubernetes.io/projected/d840464b-b486-4421-acd8-fded53601478-kube-api-access-5rmb8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.400014 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.400409 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.400586 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.422695 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rmb8\" (UniqueName: \"kubernetes.io/projected/d840464b-b486-4421-acd8-fded53601478-kube-api-access-5rmb8\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.444181 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:19 crc kubenswrapper[4829]: I1002 07:28:19.895674 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm"] Oct 02 07:28:19 crc kubenswrapper[4829]: W1002 07:28:19.904750 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd840464b_b486_4421_acd8_fded53601478.slice/crio-4feac33d8916caf0627e5e8b15143d96480a132432817f42ca26c188d3f38896 WatchSource:0}: Error finding container 4feac33d8916caf0627e5e8b15143d96480a132432817f42ca26c188d3f38896: Status 404 returned error can't find the container with id 4feac33d8916caf0627e5e8b15143d96480a132432817f42ca26c188d3f38896 Oct 02 07:28:20 crc kubenswrapper[4829]: I1002 07:28:20.486121 4829 generic.go:334] "Generic (PLEG): container finished" podID="d840464b-b486-4421-acd8-fded53601478" containerID="fd193a4a5a54d1c619998c064c6e3c02534e7be31c20b1fb22a8a6017057a77e" exitCode=0 Oct 02 07:28:20 crc kubenswrapper[4829]: I1002 07:28:20.486191 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" event={"ID":"d840464b-b486-4421-acd8-fded53601478","Type":"ContainerDied","Data":"fd193a4a5a54d1c619998c064c6e3c02534e7be31c20b1fb22a8a6017057a77e"} Oct 02 07:28:20 crc kubenswrapper[4829]: I1002 07:28:20.486314 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" event={"ID":"d840464b-b486-4421-acd8-fded53601478","Type":"ContainerStarted","Data":"4feac33d8916caf0627e5e8b15143d96480a132432817f42ca26c188d3f38896"} Oct 02 07:28:22 crc kubenswrapper[4829]: I1002 07:28:22.510502 4829 generic.go:334] "Generic (PLEG): container finished" podID="d840464b-b486-4421-acd8-fded53601478" containerID="ba043c26ec0d6e36425da1fd5f0ae57eb0241454b0b8b1d3296a3f193ec5d5da" exitCode=0 Oct 02 07:28:22 crc kubenswrapper[4829]: I1002 07:28:22.510887 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" event={"ID":"d840464b-b486-4421-acd8-fded53601478","Type":"ContainerDied","Data":"ba043c26ec0d6e36425da1fd5f0ae57eb0241454b0b8b1d3296a3f193ec5d5da"} Oct 02 07:28:23 crc kubenswrapper[4829]: I1002 07:28:23.519759 4829 generic.go:334] "Generic (PLEG): container finished" podID="d840464b-b486-4421-acd8-fded53601478" containerID="407aa70a578804bb7a7750968831f1f1d1a54fc2227ca611b3e83a4f84a5d130" exitCode=0 Oct 02 07:28:23 crc kubenswrapper[4829]: I1002 07:28:23.519825 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" event={"ID":"d840464b-b486-4421-acd8-fded53601478","Type":"ContainerDied","Data":"407aa70a578804bb7a7750968831f1f1d1a54fc2227ca611b3e83a4f84a5d130"} Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.824342 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.878257 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-bundle\") pod \"d840464b-b486-4421-acd8-fded53601478\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.878434 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-util\") pod \"d840464b-b486-4421-acd8-fded53601478\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.878541 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rmb8\" (UniqueName: \"kubernetes.io/projected/d840464b-b486-4421-acd8-fded53601478-kube-api-access-5rmb8\") pod \"d840464b-b486-4421-acd8-fded53601478\" (UID: \"d840464b-b486-4421-acd8-fded53601478\") " Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.879598 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-bundle" (OuterVolumeSpecName: "bundle") pod "d840464b-b486-4421-acd8-fded53601478" (UID: "d840464b-b486-4421-acd8-fded53601478"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.893458 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d840464b-b486-4421-acd8-fded53601478-kube-api-access-5rmb8" (OuterVolumeSpecName: "kube-api-access-5rmb8") pod "d840464b-b486-4421-acd8-fded53601478" (UID: "d840464b-b486-4421-acd8-fded53601478"). InnerVolumeSpecName "kube-api-access-5rmb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.912039 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-util" (OuterVolumeSpecName: "util") pod "d840464b-b486-4421-acd8-fded53601478" (UID: "d840464b-b486-4421-acd8-fded53601478"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.980521 4829 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.980578 4829 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d840464b-b486-4421-acd8-fded53601478-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:24 crc kubenswrapper[4829]: I1002 07:28:24.980597 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rmb8\" (UniqueName: \"kubernetes.io/projected/d840464b-b486-4421-acd8-fded53601478-kube-api-access-5rmb8\") on node \"crc\" DevicePath \"\"" Oct 02 07:28:25 crc kubenswrapper[4829]: I1002 07:28:25.534813 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" event={"ID":"d840464b-b486-4421-acd8-fded53601478","Type":"ContainerDied","Data":"4feac33d8916caf0627e5e8b15143d96480a132432817f42ca26c188d3f38896"} Oct 02 07:28:25 crc kubenswrapper[4829]: I1002 07:28:25.534872 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4feac33d8916caf0627e5e8b15143d96480a132432817f42ca26c188d3f38896" Oct 02 07:28:25 crc kubenswrapper[4829]: I1002 07:28:25.534884 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.020242 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld"] Oct 02 07:28:28 crc kubenswrapper[4829]: E1002 07:28:28.020760 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="extract" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.020777 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="extract" Oct 02 07:28:28 crc kubenswrapper[4829]: E1002 07:28:28.020792 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="util" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.020799 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="util" Oct 02 07:28:28 crc kubenswrapper[4829]: E1002 07:28:28.020820 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="pull" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.020828 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="pull" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.020944 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d840464b-b486-4421-acd8-fded53601478" containerName="extract" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.021419 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.023747 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.023950 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.025285 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-fhn58" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.039373 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld"] Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.131950 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bkz\" (UniqueName: \"kubernetes.io/projected/7bc8f7d2-12dc-43a9-8cae-872d39a4defc-kube-api-access-d5bkz\") pod \"nmstate-operator-858ddd8f98-hl4ld\" (UID: \"7bc8f7d2-12dc-43a9-8cae-872d39a4defc\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.233700 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bkz\" (UniqueName: \"kubernetes.io/projected/7bc8f7d2-12dc-43a9-8cae-872d39a4defc-kube-api-access-d5bkz\") pod \"nmstate-operator-858ddd8f98-hl4ld\" (UID: \"7bc8f7d2-12dc-43a9-8cae-872d39a4defc\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.251084 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bkz\" (UniqueName: \"kubernetes.io/projected/7bc8f7d2-12dc-43a9-8cae-872d39a4defc-kube-api-access-d5bkz\") pod \"nmstate-operator-858ddd8f98-hl4ld\" (UID: \"7bc8f7d2-12dc-43a9-8cae-872d39a4defc\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.343770 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" Oct 02 07:28:28 crc kubenswrapper[4829]: I1002 07:28:28.624702 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld"] Oct 02 07:28:29 crc kubenswrapper[4829]: I1002 07:28:29.563604 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" event={"ID":"7bc8f7d2-12dc-43a9-8cae-872d39a4defc","Type":"ContainerStarted","Data":"76341fc942f4d70b4427dfda2947e0527bc00a999390decb79a4718e4dcdf4a4"} Oct 02 07:28:31 crc kubenswrapper[4829]: I1002 07:28:31.574498 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" event={"ID":"7bc8f7d2-12dc-43a9-8cae-872d39a4defc","Type":"ContainerStarted","Data":"e4df2cd6b1308f7659ffe0f0a40c58bc10d76f9076166b53219c451c8a550fd9"} Oct 02 07:28:31 crc kubenswrapper[4829]: I1002 07:28:31.598654 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-hl4ld" podStartSLOduration=2.381950686 podStartE2EDuration="4.598623989s" podCreationTimestamp="2025-10-02 07:28:27 +0000 UTC" firstStartedPulling="2025-10-02 07:28:28.631106143 +0000 UTC m=+699.970754548" lastFinishedPulling="2025-10-02 07:28:30.847779446 +0000 UTC m=+702.187427851" observedRunningTime="2025-10-02 07:28:31.595140182 +0000 UTC m=+702.934788587" watchObservedRunningTime="2025-10-02 07:28:31.598623989 +0000 UTC m=+702.938272434" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.669680 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.670644 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.673299 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-l8qk9" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.681507 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.685515 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.686095 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtgsw\" (UniqueName: \"kubernetes.io/projected/3795be47-c66b-4804-b130-33a73c5e57ce-kube-api-access-dtgsw\") pod \"nmstate-metrics-fdff9cb8d-8rhft\" (UID: \"3795be47-c66b-4804-b130-33a73c5e57ce\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.686281 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.719022 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.727421 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.746361 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kgg6f"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.747398 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.786824 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-nmstate-lock\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.786888 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzhfm\" (UniqueName: \"kubernetes.io/projected/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-kube-api-access-nzhfm\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.786926 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7e10a126-c321-4d0b-9757-080946eddbba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-6q26v\" (UID: \"7e10a126-c321-4d0b-9757-080946eddbba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.786946 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-ovs-socket\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.786976 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28lnn\" (UniqueName: \"kubernetes.io/projected/7e10a126-c321-4d0b-9757-080946eddbba-kube-api-access-28lnn\") pod \"nmstate-webhook-6cdbc54649-6q26v\" (UID: \"7e10a126-c321-4d0b-9757-080946eddbba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.787018 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtgsw\" (UniqueName: \"kubernetes.io/projected/3795be47-c66b-4804-b130-33a73c5e57ce-kube-api-access-dtgsw\") pod \"nmstate-metrics-fdff9cb8d-8rhft\" (UID: \"3795be47-c66b-4804-b130-33a73c5e57ce\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.787038 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-dbus-socket\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.808296 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtgsw\" (UniqueName: \"kubernetes.io/projected/3795be47-c66b-4804-b130-33a73c5e57ce-kube-api-access-dtgsw\") pod \"nmstate-metrics-fdff9cb8d-8rhft\" (UID: \"3795be47-c66b-4804-b130-33a73c5e57ce\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.850592 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.851346 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.854309 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.854364 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-ncdck" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.854305 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.862778 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr"] Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.887872 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzhfm\" (UniqueName: \"kubernetes.io/projected/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-kube-api-access-nzhfm\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.887942 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7e10a126-c321-4d0b-9757-080946eddbba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-6q26v\" (UID: \"7e10a126-c321-4d0b-9757-080946eddbba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.887972 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtgmk\" (UniqueName: \"kubernetes.io/projected/d9eea7b1-65ab-4699-ad35-15aad5a523f9-kube-api-access-dtgmk\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.888001 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-ovs-socket\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.888034 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28lnn\" (UniqueName: \"kubernetes.io/projected/7e10a126-c321-4d0b-9757-080946eddbba-kube-api-access-28lnn\") pod \"nmstate-webhook-6cdbc54649-6q26v\" (UID: \"7e10a126-c321-4d0b-9757-080946eddbba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.888062 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9eea7b1-65ab-4699-ad35-15aad5a523f9-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.888108 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-dbus-socket\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.888157 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-nmstate-lock\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.888192 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9eea7b1-65ab-4699-ad35-15aad5a523f9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.889005 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-ovs-socket\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.889051 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-nmstate-lock\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.889255 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-dbus-socket\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.892021 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/7e10a126-c321-4d0b-9757-080946eddbba-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-6q26v\" (UID: \"7e10a126-c321-4d0b-9757-080946eddbba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.901730 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzhfm\" (UniqueName: \"kubernetes.io/projected/6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e-kube-api-access-nzhfm\") pod \"nmstate-handler-kgg6f\" (UID: \"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e\") " pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.911099 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28lnn\" (UniqueName: \"kubernetes.io/projected/7e10a126-c321-4d0b-9757-080946eddbba-kube-api-access-28lnn\") pod \"nmstate-webhook-6cdbc54649-6q26v\" (UID: \"7e10a126-c321-4d0b-9757-080946eddbba\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.989508 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9eea7b1-65ab-4699-ad35-15aad5a523f9-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.989619 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9eea7b1-65ab-4699-ad35-15aad5a523f9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.990029 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtgmk\" (UniqueName: \"kubernetes.io/projected/d9eea7b1-65ab-4699-ad35-15aad5a523f9-kube-api-access-dtgmk\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.990209 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9eea7b1-65ab-4699-ad35-15aad5a523f9-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:32 crc kubenswrapper[4829]: I1002 07:28:32.994966 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9eea7b1-65ab-4699-ad35-15aad5a523f9-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.004914 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtgmk\" (UniqueName: \"kubernetes.io/projected/d9eea7b1-65ab-4699-ad35-15aad5a523f9-kube-api-access-dtgmk\") pod \"nmstate-console-plugin-6b874cbd85-t5rvr\" (UID: \"d9eea7b1-65ab-4699-ad35-15aad5a523f9\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.026672 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.045013 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.055146 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-799ccdbf7b-4brqq"] Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.055949 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.066656 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.082284 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-799ccdbf7b-4brqq"] Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097429 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ecda265f-de2f-4ac7-966d-c945f8307066-console-serving-cert\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097488 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klgg7\" (UniqueName: \"kubernetes.io/projected/ecda265f-de2f-4ac7-966d-c945f8307066-kube-api-access-klgg7\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097513 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-service-ca\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097559 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-trusted-ca-bundle\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097591 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-console-config\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097663 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-oauth-serving-cert\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.097695 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ecda265f-de2f-4ac7-966d-c945f8307066-console-oauth-config\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.164429 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199122 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klgg7\" (UniqueName: \"kubernetes.io/projected/ecda265f-de2f-4ac7-966d-c945f8307066-kube-api-access-klgg7\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199158 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-service-ca\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199188 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-trusted-ca-bundle\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199209 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-console-config\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199293 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-oauth-serving-cert\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199317 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ecda265f-de2f-4ac7-966d-c945f8307066-console-oauth-config\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.199349 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ecda265f-de2f-4ac7-966d-c945f8307066-console-serving-cert\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.200923 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-trusted-ca-bundle\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.201754 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-service-ca\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.202277 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-oauth-serving-cert\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.202761 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ecda265f-de2f-4ac7-966d-c945f8307066-console-config\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.205646 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ecda265f-de2f-4ac7-966d-c945f8307066-console-oauth-config\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.206374 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ecda265f-de2f-4ac7-966d-c945f8307066-console-serving-cert\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.229044 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klgg7\" (UniqueName: \"kubernetes.io/projected/ecda265f-de2f-4ac7-966d-c945f8307066-kube-api-access-klgg7\") pod \"console-799ccdbf7b-4brqq\" (UID: \"ecda265f-de2f-4ac7-966d-c945f8307066\") " pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.308509 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft"] Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.371409 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr"] Oct 02 07:28:33 crc kubenswrapper[4829]: W1002 07:28:33.374546 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9eea7b1_65ab_4699_ad35_15aad5a523f9.slice/crio-d42ea0a0813ae7a3f808f9375686c47e75a8dfda2c781cf1a4e712059d3dcd70 WatchSource:0}: Error finding container d42ea0a0813ae7a3f808f9375686c47e75a8dfda2c781cf1a4e712059d3dcd70: Status 404 returned error can't find the container with id d42ea0a0813ae7a3f808f9375686c47e75a8dfda2c781cf1a4e712059d3dcd70 Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.376091 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.559727 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-799ccdbf7b-4brqq"] Oct 02 07:28:33 crc kubenswrapper[4829]: W1002 07:28:33.564990 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecda265f_de2f_4ac7_966d_c945f8307066.slice/crio-02f4ed7a262a5ce5a6693b73409d92bcd4d20759417ae23f2a5a9a1f44c73b71 WatchSource:0}: Error finding container 02f4ed7a262a5ce5a6693b73409d92bcd4d20759417ae23f2a5a9a1f44c73b71: Status 404 returned error can't find the container with id 02f4ed7a262a5ce5a6693b73409d92bcd4d20759417ae23f2a5a9a1f44c73b71 Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.584885 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kgg6f" event={"ID":"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e","Type":"ContainerStarted","Data":"8b6a68c2e26dbcc124baadfeaeade49aeadbf78b2c832531826232271e6496e9"} Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.585942 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" event={"ID":"d9eea7b1-65ab-4699-ad35-15aad5a523f9","Type":"ContainerStarted","Data":"d42ea0a0813ae7a3f808f9375686c47e75a8dfda2c781cf1a4e712059d3dcd70"} Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.587567 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" event={"ID":"3795be47-c66b-4804-b130-33a73c5e57ce","Type":"ContainerStarted","Data":"3ee20e063a58cf631142707789a9e8596188994f6471588f713347ecd65f9c19"} Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.588505 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-799ccdbf7b-4brqq" event={"ID":"ecda265f-de2f-4ac7-966d-c945f8307066","Type":"ContainerStarted","Data":"02f4ed7a262a5ce5a6693b73409d92bcd4d20759417ae23f2a5a9a1f44c73b71"} Oct 02 07:28:33 crc kubenswrapper[4829]: I1002 07:28:33.614443 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v"] Oct 02 07:28:33 crc kubenswrapper[4829]: W1002 07:28:33.620427 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e10a126_c321_4d0b_9757_080946eddbba.slice/crio-dddc8a92c6be43918c17f31c9e447ebea2f5e7c98a21751e8a671d122bd97293 WatchSource:0}: Error finding container dddc8a92c6be43918c17f31c9e447ebea2f5e7c98a21751e8a671d122bd97293: Status 404 returned error can't find the container with id dddc8a92c6be43918c17f31c9e447ebea2f5e7c98a21751e8a671d122bd97293 Oct 02 07:28:34 crc kubenswrapper[4829]: I1002 07:28:34.596630 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-799ccdbf7b-4brqq" event={"ID":"ecda265f-de2f-4ac7-966d-c945f8307066","Type":"ContainerStarted","Data":"270483ab1dc1103ce977912b2015fa054f34171a12f0af2a6055bc97b8244fcd"} Oct 02 07:28:34 crc kubenswrapper[4829]: I1002 07:28:34.597560 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" event={"ID":"7e10a126-c321-4d0b-9757-080946eddbba","Type":"ContainerStarted","Data":"dddc8a92c6be43918c17f31c9e447ebea2f5e7c98a21751e8a671d122bd97293"} Oct 02 07:28:34 crc kubenswrapper[4829]: I1002 07:28:34.617726 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-799ccdbf7b-4brqq" podStartSLOduration=1.6177042259999999 podStartE2EDuration="1.617704226s" podCreationTimestamp="2025-10-02 07:28:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:28:34.612560007 +0000 UTC m=+705.952208432" watchObservedRunningTime="2025-10-02 07:28:34.617704226 +0000 UTC m=+705.957352641" Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.612196 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kgg6f" event={"ID":"6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e","Type":"ContainerStarted","Data":"629be71ee8c5bd9eecce8e17a0bda28807012071ba5c9f70fb16233c09589514"} Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.612877 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.615330 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" event={"ID":"d9eea7b1-65ab-4699-ad35-15aad5a523f9","Type":"ContainerStarted","Data":"21097b4acfcf7cf252de8a0ed7573de2c9fcf902ee2d01919e2913830d60d84d"} Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.617627 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" event={"ID":"7e10a126-c321-4d0b-9757-080946eddbba","Type":"ContainerStarted","Data":"61610e6aba26f8e12da1cc5900d628f2a902a0cba5eb172ed0aec147729973ca"} Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.617778 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.620955 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" event={"ID":"3795be47-c66b-4804-b130-33a73c5e57ce","Type":"ContainerStarted","Data":"14c971289b4f2e197fad5631d7c80949655d8ecd391081bb2210ba2b964c4644"} Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.635017 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kgg6f" podStartSLOduration=1.9619982519999999 podStartE2EDuration="4.63498927s" podCreationTimestamp="2025-10-02 07:28:32 +0000 UTC" firstStartedPulling="2025-10-02 07:28:33.109197939 +0000 UTC m=+704.448846344" lastFinishedPulling="2025-10-02 07:28:35.782188957 +0000 UTC m=+707.121837362" observedRunningTime="2025-10-02 07:28:36.628753828 +0000 UTC m=+707.968402303" watchObservedRunningTime="2025-10-02 07:28:36.63498927 +0000 UTC m=+707.974637705" Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.649424 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" podStartSLOduration=2.485355179 podStartE2EDuration="4.649405954s" podCreationTimestamp="2025-10-02 07:28:32 +0000 UTC" firstStartedPulling="2025-10-02 07:28:33.622345462 +0000 UTC m=+704.961993877" lastFinishedPulling="2025-10-02 07:28:35.786396227 +0000 UTC m=+707.126044652" observedRunningTime="2025-10-02 07:28:36.649147966 +0000 UTC m=+707.988796401" watchObservedRunningTime="2025-10-02 07:28:36.649405954 +0000 UTC m=+707.989054349" Oct 02 07:28:36 crc kubenswrapper[4829]: I1002 07:28:36.667164 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-t5rvr" podStartSLOduration=2.262014392 podStartE2EDuration="4.66712964s" podCreationTimestamp="2025-10-02 07:28:32 +0000 UTC" firstStartedPulling="2025-10-02 07:28:33.376715068 +0000 UTC m=+704.716363473" lastFinishedPulling="2025-10-02 07:28:35.781830326 +0000 UTC m=+707.121478721" observedRunningTime="2025-10-02 07:28:36.663390355 +0000 UTC m=+708.003038820" watchObservedRunningTime="2025-10-02 07:28:36.66712964 +0000 UTC m=+708.006778095" Oct 02 07:28:39 crc kubenswrapper[4829]: I1002 07:28:39.653004 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" event={"ID":"3795be47-c66b-4804-b130-33a73c5e57ce","Type":"ContainerStarted","Data":"4e35d71f504d0b1a5baf3087ed3131a2b78c44de0c9ee8d4d3dfa9b871d764da"} Oct 02 07:28:39 crc kubenswrapper[4829]: I1002 07:28:39.677597 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8rhft" podStartSLOduration=1.8875338080000001 podStartE2EDuration="7.67757585s" podCreationTimestamp="2025-10-02 07:28:32 +0000 UTC" firstStartedPulling="2025-10-02 07:28:33.316294956 +0000 UTC m=+704.655943361" lastFinishedPulling="2025-10-02 07:28:39.106336958 +0000 UTC m=+710.445985403" observedRunningTime="2025-10-02 07:28:39.676679802 +0000 UTC m=+711.016328237" watchObservedRunningTime="2025-10-02 07:28:39.67757585 +0000 UTC m=+711.017224265" Oct 02 07:28:43 crc kubenswrapper[4829]: I1002 07:28:43.103575 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kgg6f" Oct 02 07:28:43 crc kubenswrapper[4829]: I1002 07:28:43.376371 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:43 crc kubenswrapper[4829]: I1002 07:28:43.376446 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:43 crc kubenswrapper[4829]: I1002 07:28:43.383710 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:43 crc kubenswrapper[4829]: I1002 07:28:43.687014 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-799ccdbf7b-4brqq" Oct 02 07:28:43 crc kubenswrapper[4829]: I1002 07:28:43.757001 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2tvsb"] Oct 02 07:28:53 crc kubenswrapper[4829]: I1002 07:28:53.053806 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-6q26v" Oct 02 07:28:55 crc kubenswrapper[4829]: I1002 07:28:55.329422 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:28:55 crc kubenswrapper[4829]: I1002 07:28:55.329848 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.242418 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7"] Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.244158 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.246712 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.264732 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7"] Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.409891 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.409994 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klzg6\" (UniqueName: \"kubernetes.io/projected/036e78a8-b2cb-4210-bbfd-dedba601b30a-kube-api-access-klzg6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.410022 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.512018 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klzg6\" (UniqueName: \"kubernetes.io/projected/036e78a8-b2cb-4210-bbfd-dedba601b30a-kube-api-access-klzg6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.512104 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.512152 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.513027 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.514023 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.541037 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klzg6\" (UniqueName: \"kubernetes.io/projected/036e78a8-b2cb-4210-bbfd-dedba601b30a-kube-api-access-klzg6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.562784 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.809208 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7"] Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.822160 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-2tvsb" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerName="console" containerID="cri-o://27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c" gracePeriod=15 Oct 02 07:29:08 crc kubenswrapper[4829]: I1002 07:29:08.866859 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" event={"ID":"036e78a8-b2cb-4210-bbfd-dedba601b30a","Type":"ContainerStarted","Data":"24588166a96608fd3b0a50ebc7a505177b34f63a81f3afae6805431a97e3442e"} Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.203895 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2tvsb_4c7291f2-090f-4ab6-bcd5-203fee2bbce1/console/0.log" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.203959 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322718 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-serving-cert\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322815 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-config\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322854 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-trusted-ca-bundle\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322872 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-oauth-serving-cert\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322899 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-service-ca\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322925 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-oauth-config\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.322972 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v85pb\" (UniqueName: \"kubernetes.io/projected/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-kube-api-access-v85pb\") pod \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\" (UID: \"4c7291f2-090f-4ab6-bcd5-203fee2bbce1\") " Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.324561 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.324623 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-service-ca" (OuterVolumeSpecName: "service-ca") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.324633 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-config" (OuterVolumeSpecName: "console-config") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.324656 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.328521 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.328808 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-kube-api-access-v85pb" (OuterVolumeSpecName: "kube-api-access-v85pb") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "kube-api-access-v85pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.328807 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4c7291f2-090f-4ab6-bcd5-203fee2bbce1" (UID: "4c7291f2-090f-4ab6-bcd5-203fee2bbce1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424091 4829 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424147 4829 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424164 4829 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424178 4829 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424189 4829 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424200 4829 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.424212 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v85pb\" (UniqueName: \"kubernetes.io/projected/4c7291f2-090f-4ab6-bcd5-203fee2bbce1-kube-api-access-v85pb\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.875215 4829 generic.go:334] "Generic (PLEG): container finished" podID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerID="86998290a2153334a70e49e4f4e690e8cc4b596c368196b08ff5596cbdac6196" exitCode=0 Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.875532 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" event={"ID":"036e78a8-b2cb-4210-bbfd-dedba601b30a","Type":"ContainerDied","Data":"86998290a2153334a70e49e4f4e690e8cc4b596c368196b08ff5596cbdac6196"} Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.880389 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2tvsb_4c7291f2-090f-4ab6-bcd5-203fee2bbce1/console/0.log" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.880598 4829 generic.go:334] "Generic (PLEG): container finished" podID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerID="27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c" exitCode=2 Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.880667 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2tvsb" event={"ID":"4c7291f2-090f-4ab6-bcd5-203fee2bbce1","Type":"ContainerDied","Data":"27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c"} Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.881993 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2tvsb" event={"ID":"4c7291f2-090f-4ab6-bcd5-203fee2bbce1","Type":"ContainerDied","Data":"5cfcc4ee06baacdcb73ca3effabc7e025cfab177e1da0eb52631f12fb70752b6"} Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.880716 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2tvsb" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.882356 4829 scope.go:117] "RemoveContainer" containerID="27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.904771 4829 scope.go:117] "RemoveContainer" containerID="27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c" Oct 02 07:29:09 crc kubenswrapper[4829]: E1002 07:29:09.905419 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c\": container with ID starting with 27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c not found: ID does not exist" containerID="27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.905469 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c"} err="failed to get container status \"27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c\": rpc error: code = NotFound desc = could not find container \"27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c\": container with ID starting with 27f7fc12522af70b42b7c085ce286aeaea216122e7ef71544fd0dc8699419b9c not found: ID does not exist" Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.924388 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2tvsb"] Oct 02 07:29:09 crc kubenswrapper[4829]: I1002 07:29:09.929975 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-2tvsb"] Oct 02 07:29:11 crc kubenswrapper[4829]: I1002 07:29:11.468889 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" path="/var/lib/kubelet/pods/4c7291f2-090f-4ab6-bcd5-203fee2bbce1/volumes" Oct 02 07:29:11 crc kubenswrapper[4829]: I1002 07:29:11.896562 4829 generic.go:334] "Generic (PLEG): container finished" podID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerID="3820ce58427790b5e0458e9023a029f07829d27d353366291bebfd63f4976457" exitCode=0 Oct 02 07:29:11 crc kubenswrapper[4829]: I1002 07:29:11.896604 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" event={"ID":"036e78a8-b2cb-4210-bbfd-dedba601b30a","Type":"ContainerDied","Data":"3820ce58427790b5e0458e9023a029f07829d27d353366291bebfd63f4976457"} Oct 02 07:29:12 crc kubenswrapper[4829]: I1002 07:29:12.906372 4829 generic.go:334] "Generic (PLEG): container finished" podID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerID="f4f981ac711602977b1a653ef5a118fc18f2da24d58233bc727701db589c78d3" exitCode=0 Oct 02 07:29:12 crc kubenswrapper[4829]: I1002 07:29:12.906486 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" event={"ID":"036e78a8-b2cb-4210-bbfd-dedba601b30a","Type":"ContainerDied","Data":"f4f981ac711602977b1a653ef5a118fc18f2da24d58233bc727701db589c78d3"} Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.124634 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.288721 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-bundle\") pod \"036e78a8-b2cb-4210-bbfd-dedba601b30a\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.288772 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klzg6\" (UniqueName: \"kubernetes.io/projected/036e78a8-b2cb-4210-bbfd-dedba601b30a-kube-api-access-klzg6\") pod \"036e78a8-b2cb-4210-bbfd-dedba601b30a\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.288831 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-util\") pod \"036e78a8-b2cb-4210-bbfd-dedba601b30a\" (UID: \"036e78a8-b2cb-4210-bbfd-dedba601b30a\") " Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.290724 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-bundle" (OuterVolumeSpecName: "bundle") pod "036e78a8-b2cb-4210-bbfd-dedba601b30a" (UID: "036e78a8-b2cb-4210-bbfd-dedba601b30a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.294631 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036e78a8-b2cb-4210-bbfd-dedba601b30a-kube-api-access-klzg6" (OuterVolumeSpecName: "kube-api-access-klzg6") pod "036e78a8-b2cb-4210-bbfd-dedba601b30a" (UID: "036e78a8-b2cb-4210-bbfd-dedba601b30a"). InnerVolumeSpecName "kube-api-access-klzg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.320539 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-util" (OuterVolumeSpecName: "util") pod "036e78a8-b2cb-4210-bbfd-dedba601b30a" (UID: "036e78a8-b2cb-4210-bbfd-dedba601b30a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.390481 4829 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.390537 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klzg6\" (UniqueName: \"kubernetes.io/projected/036e78a8-b2cb-4210-bbfd-dedba601b30a-kube-api-access-klzg6\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.390559 4829 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/036e78a8-b2cb-4210-bbfd-dedba601b30a-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.926199 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" event={"ID":"036e78a8-b2cb-4210-bbfd-dedba601b30a","Type":"ContainerDied","Data":"24588166a96608fd3b0a50ebc7a505177b34f63a81f3afae6805431a97e3442e"} Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.926280 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24588166a96608fd3b0a50ebc7a505177b34f63a81f3afae6805431a97e3442e" Oct 02 07:29:14 crc kubenswrapper[4829]: I1002 07:29:14.926289 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.016559 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsrrk"] Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.017141 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" podUID="8706882f-e939-4ac9-b145-527cb52f9294" containerName="controller-manager" containerID="cri-o://3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883" gracePeriod=30 Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.020430 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8"] Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.020629 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" podUID="6e908e8d-4252-422b-9d40-f38ef0766dde" containerName="route-controller-manager" containerID="cri-o://f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5" gracePeriod=30 Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.537105 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.540812 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637616 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8706882f-e939-4ac9-b145-527cb52f9294-serving-cert\") pod \"8706882f-e939-4ac9-b145-527cb52f9294\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637690 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-proxy-ca-bundles\") pod \"8706882f-e939-4ac9-b145-527cb52f9294\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637713 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j5n7\" (UniqueName: \"kubernetes.io/projected/8706882f-e939-4ac9-b145-527cb52f9294-kube-api-access-7j5n7\") pod \"8706882f-e939-4ac9-b145-527cb52f9294\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637777 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e908e8d-4252-422b-9d40-f38ef0766dde-serving-cert\") pod \"6e908e8d-4252-422b-9d40-f38ef0766dde\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637824 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-config\") pod \"8706882f-e939-4ac9-b145-527cb52f9294\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637856 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s8tg\" (UniqueName: \"kubernetes.io/projected/6e908e8d-4252-422b-9d40-f38ef0766dde-kube-api-access-4s8tg\") pod \"6e908e8d-4252-422b-9d40-f38ef0766dde\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637872 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-client-ca\") pod \"6e908e8d-4252-422b-9d40-f38ef0766dde\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637890 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-config\") pod \"6e908e8d-4252-422b-9d40-f38ef0766dde\" (UID: \"6e908e8d-4252-422b-9d40-f38ef0766dde\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.637917 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-client-ca\") pod \"8706882f-e939-4ac9-b145-527cb52f9294\" (UID: \"8706882f-e939-4ac9-b145-527cb52f9294\") " Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.638417 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-client-ca" (OuterVolumeSpecName: "client-ca") pod "6e908e8d-4252-422b-9d40-f38ef0766dde" (UID: "6e908e8d-4252-422b-9d40-f38ef0766dde"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.638814 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-config" (OuterVolumeSpecName: "config") pod "8706882f-e939-4ac9-b145-527cb52f9294" (UID: "8706882f-e939-4ac9-b145-527cb52f9294"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.638835 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-client-ca" (OuterVolumeSpecName: "client-ca") pod "8706882f-e939-4ac9-b145-527cb52f9294" (UID: "8706882f-e939-4ac9-b145-527cb52f9294"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.638890 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-config" (OuterVolumeSpecName: "config") pod "6e908e8d-4252-422b-9d40-f38ef0766dde" (UID: "6e908e8d-4252-422b-9d40-f38ef0766dde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.639129 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8706882f-e939-4ac9-b145-527cb52f9294" (UID: "8706882f-e939-4ac9-b145-527cb52f9294"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.643759 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e908e8d-4252-422b-9d40-f38ef0766dde-kube-api-access-4s8tg" (OuterVolumeSpecName: "kube-api-access-4s8tg") pod "6e908e8d-4252-422b-9d40-f38ef0766dde" (UID: "6e908e8d-4252-422b-9d40-f38ef0766dde"). InnerVolumeSpecName "kube-api-access-4s8tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.643807 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8706882f-e939-4ac9-b145-527cb52f9294-kube-api-access-7j5n7" (OuterVolumeSpecName: "kube-api-access-7j5n7") pod "8706882f-e939-4ac9-b145-527cb52f9294" (UID: "8706882f-e939-4ac9-b145-527cb52f9294"). InnerVolumeSpecName "kube-api-access-7j5n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.643880 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e908e8d-4252-422b-9d40-f38ef0766dde-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6e908e8d-4252-422b-9d40-f38ef0766dde" (UID: "6e908e8d-4252-422b-9d40-f38ef0766dde"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.646418 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8706882f-e939-4ac9-b145-527cb52f9294-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8706882f-e939-4ac9-b145-527cb52f9294" (UID: "8706882f-e939-4ac9-b145-527cb52f9294"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739213 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739258 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s8tg\" (UniqueName: \"kubernetes.io/projected/6e908e8d-4252-422b-9d40-f38ef0766dde-kube-api-access-4s8tg\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739269 4829 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739279 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e908e8d-4252-422b-9d40-f38ef0766dde-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739287 4829 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739296 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8706882f-e939-4ac9-b145-527cb52f9294-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739304 4829 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8706882f-e939-4ac9-b145-527cb52f9294-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739313 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j5n7\" (UniqueName: \"kubernetes.io/projected/8706882f-e939-4ac9-b145-527cb52f9294-kube-api-access-7j5n7\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.739323 4829 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e908e8d-4252-422b-9d40-f38ef0766dde-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.945888 4829 generic.go:334] "Generic (PLEG): container finished" podID="8706882f-e939-4ac9-b145-527cb52f9294" containerID="3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883" exitCode=0 Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.945962 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.945978 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" event={"ID":"8706882f-e939-4ac9-b145-527cb52f9294","Type":"ContainerDied","Data":"3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883"} Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.946508 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tsrrk" event={"ID":"8706882f-e939-4ac9-b145-527cb52f9294","Type":"ContainerDied","Data":"5dba925553d464183ca35d9f1ee8b6c6b6f43691259c2555c6ffdef1c5c4de91"} Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.946527 4829 scope.go:117] "RemoveContainer" containerID="3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.947779 4829 generic.go:334] "Generic (PLEG): container finished" podID="6e908e8d-4252-422b-9d40-f38ef0766dde" containerID="f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5" exitCode=0 Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.947831 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.947841 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" event={"ID":"6e908e8d-4252-422b-9d40-f38ef0766dde","Type":"ContainerDied","Data":"f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5"} Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.947869 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8" event={"ID":"6e908e8d-4252-422b-9d40-f38ef0766dde","Type":"ContainerDied","Data":"b05012b089903b3c8e30de0b3526d9a8815550c2fd7a004db467c17249e97f24"} Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.969374 4829 scope.go:117] "RemoveContainer" containerID="3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883" Oct 02 07:29:17 crc kubenswrapper[4829]: E1002 07:29:17.969851 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883\": container with ID starting with 3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883 not found: ID does not exist" containerID="3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.969895 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883"} err="failed to get container status \"3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883\": rpc error: code = NotFound desc = could not find container \"3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883\": container with ID starting with 3b9c0c9254e92964b0b096db830f93f67973eac35274b17b2590a45627e0b883 not found: ID does not exist" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.969923 4829 scope.go:117] "RemoveContainer" containerID="f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5" Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.984474 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8"] Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.990623 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-7d9w8"] Oct 02 07:29:17 crc kubenswrapper[4829]: I1002 07:29:17.999178 4829 scope.go:117] "RemoveContainer" containerID="f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5" Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.000423 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5\": container with ID starting with f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5 not found: ID does not exist" containerID="f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.000496 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5"} err="failed to get container status \"f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5\": rpc error: code = NotFound desc = could not find container \"f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5\": container with ID starting with f2448eb45df1cb6d83e1d3db29448036983034dcec9ac55c44de126e2103fdd5 not found: ID does not exist" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.003507 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsrrk"] Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.007543 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tsrrk"] Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.282725 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d58848857-wfv8z"] Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.282913 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="pull" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.282924 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="pull" Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.282931 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e908e8d-4252-422b-9d40-f38ef0766dde" containerName="route-controller-manager" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.282937 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e908e8d-4252-422b-9d40-f38ef0766dde" containerName="route-controller-manager" Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.282947 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8706882f-e939-4ac9-b145-527cb52f9294" containerName="controller-manager" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.282953 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8706882f-e939-4ac9-b145-527cb52f9294" containerName="controller-manager" Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.282964 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="extract" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.282969 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="extract" Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.282978 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="util" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.282983 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="util" Oct 02 07:29:18 crc kubenswrapper[4829]: E1002 07:29:18.282996 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerName="console" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.283002 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerName="console" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.283094 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7291f2-090f-4ab6-bcd5-203fee2bbce1" containerName="console" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.283106 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="036e78a8-b2cb-4210-bbfd-dedba601b30a" containerName="extract" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.283115 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8706882f-e939-4ac9-b145-527cb52f9294" containerName="controller-manager" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.283124 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e908e8d-4252-422b-9d40-f38ef0766dde" containerName="route-controller-manager" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.283490 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.286839 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.286996 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.287074 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.289137 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.290652 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.298312 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.302057 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d58848857-wfv8z"] Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.318003 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.344133 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-config\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.344251 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-proxy-ca-bundles\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.344279 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2xb7\" (UniqueName: \"kubernetes.io/projected/56959900-ee62-4c7b-ba8d-b6a423aa3241-kube-api-access-r2xb7\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.344304 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-client-ca\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.344545 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56959900-ee62-4c7b-ba8d-b6a423aa3241-serving-cert\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.445940 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56959900-ee62-4c7b-ba8d-b6a423aa3241-serving-cert\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.446835 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-config\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.446894 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-proxy-ca-bundles\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.447001 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2xb7\" (UniqueName: \"kubernetes.io/projected/56959900-ee62-4c7b-ba8d-b6a423aa3241-kube-api-access-r2xb7\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.447043 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-client-ca\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.448135 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-client-ca\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.448775 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-proxy-ca-bundles\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.449645 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56959900-ee62-4c7b-ba8d-b6a423aa3241-config\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.451378 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56959900-ee62-4c7b-ba8d-b6a423aa3241-serving-cert\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.469489 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2xb7\" (UniqueName: \"kubernetes.io/projected/56959900-ee62-4c7b-ba8d-b6a423aa3241-kube-api-access-r2xb7\") pod \"controller-manager-6d58848857-wfv8z\" (UID: \"56959900-ee62-4c7b-ba8d-b6a423aa3241\") " pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.603661 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.885696 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d58848857-wfv8z"] Oct 02 07:29:18 crc kubenswrapper[4829]: I1002 07:29:18.958055 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" event={"ID":"56959900-ee62-4c7b-ba8d-b6a423aa3241","Type":"ContainerStarted","Data":"6b82b3272022ddd290a21080d0ffa3692d7fa8235490c76661e7844b2502cc65"} Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.287553 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx"] Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.288672 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.293459 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.293773 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.293899 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.294141 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.294254 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.303070 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx"] Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.312189 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.358135 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d719597-ad1b-47ff-a3f8-66ed125371ef-serving-cert\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.358200 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d719597-ad1b-47ff-a3f8-66ed125371ef-client-ca\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.358247 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x27ns\" (UniqueName: \"kubernetes.io/projected/6d719597-ad1b-47ff-a3f8-66ed125371ef-kube-api-access-x27ns\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.358468 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d719597-ad1b-47ff-a3f8-66ed125371ef-config\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.459748 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x27ns\" (UniqueName: \"kubernetes.io/projected/6d719597-ad1b-47ff-a3f8-66ed125371ef-kube-api-access-x27ns\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.459831 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d719597-ad1b-47ff-a3f8-66ed125371ef-config\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.459870 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d719597-ad1b-47ff-a3f8-66ed125371ef-serving-cert\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.459888 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d719597-ad1b-47ff-a3f8-66ed125371ef-client-ca\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.460745 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6d719597-ad1b-47ff-a3f8-66ed125371ef-client-ca\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.461070 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d719597-ad1b-47ff-a3f8-66ed125371ef-config\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.470152 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6d719597-ad1b-47ff-a3f8-66ed125371ef-serving-cert\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.477476 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e908e8d-4252-422b-9d40-f38ef0766dde" path="/var/lib/kubelet/pods/6e908e8d-4252-422b-9d40-f38ef0766dde/volumes" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.478305 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8706882f-e939-4ac9-b145-527cb52f9294" path="/var/lib/kubelet/pods/8706882f-e939-4ac9-b145-527cb52f9294/volumes" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.501902 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x27ns\" (UniqueName: \"kubernetes.io/projected/6d719597-ad1b-47ff-a3f8-66ed125371ef-kube-api-access-x27ns\") pod \"route-controller-manager-779d575dfc-mfwmx\" (UID: \"6d719597-ad1b-47ff-a3f8-66ed125371ef\") " pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.604586 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.965710 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" event={"ID":"56959900-ee62-4c7b-ba8d-b6a423aa3241","Type":"ContainerStarted","Data":"6e923262007e411d2aa8999327ac69190de23df93f44ec58ca68e684fd40521b"} Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.965990 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.971619 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" Oct 02 07:29:19 crc kubenswrapper[4829]: I1002 07:29:19.987465 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d58848857-wfv8z" podStartSLOduration=2.987446396 podStartE2EDuration="2.987446396s" podCreationTimestamp="2025-10-02 07:29:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:29:19.984001233 +0000 UTC m=+751.323649638" watchObservedRunningTime="2025-10-02 07:29:19.987446396 +0000 UTC m=+751.327094801" Oct 02 07:29:20 crc kubenswrapper[4829]: I1002 07:29:20.011738 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx"] Oct 02 07:29:20 crc kubenswrapper[4829]: W1002 07:29:20.019192 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d719597_ad1b_47ff_a3f8_66ed125371ef.slice/crio-0cbdd4eb4bc9f7b83acdbdd300471a296174a9cfca9420b68a70aaef81cc57d7 WatchSource:0}: Error finding container 0cbdd4eb4bc9f7b83acdbdd300471a296174a9cfca9420b68a70aaef81cc57d7: Status 404 returned error can't find the container with id 0cbdd4eb4bc9f7b83acdbdd300471a296174a9cfca9420b68a70aaef81cc57d7 Oct 02 07:29:20 crc kubenswrapper[4829]: I1002 07:29:20.971522 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" event={"ID":"6d719597-ad1b-47ff-a3f8-66ed125371ef","Type":"ContainerStarted","Data":"28d139159684365eb73d3a4467221cbbfc983ba9b7aa7b7999e6027eb0875daa"} Oct 02 07:29:20 crc kubenswrapper[4829]: I1002 07:29:20.971875 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" event={"ID":"6d719597-ad1b-47ff-a3f8-66ed125371ef","Type":"ContainerStarted","Data":"0cbdd4eb4bc9f7b83acdbdd300471a296174a9cfca9420b68a70aaef81cc57d7"} Oct 02 07:29:20 crc kubenswrapper[4829]: I1002 07:29:20.987020 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" podStartSLOduration=3.986999692 podStartE2EDuration="3.986999692s" podCreationTimestamp="2025-10-02 07:29:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:29:20.985650403 +0000 UTC m=+752.325298838" watchObservedRunningTime="2025-10-02 07:29:20.986999692 +0000 UTC m=+752.326648107" Oct 02 07:29:21 crc kubenswrapper[4829]: I1002 07:29:21.977943 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:21 crc kubenswrapper[4829]: I1002 07:29:21.982652 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-779d575dfc-mfwmx" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.191785 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq"] Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.192990 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.201540 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202122 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdf6m\" (UniqueName: \"kubernetes.io/projected/65093a97-cc2b-4983-a106-60b3af7783ca-kube-api-access-hdf6m\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202176 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65093a97-cc2b-4983-a106-60b3af7783ca-webhook-cert\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202204 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65093a97-cc2b-4983-a106-60b3af7783ca-apiservice-cert\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202524 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202539 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202708 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-wdzvb" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.202710 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.208962 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq"] Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.302702 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdf6m\" (UniqueName: \"kubernetes.io/projected/65093a97-cc2b-4983-a106-60b3af7783ca-kube-api-access-hdf6m\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.302763 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65093a97-cc2b-4983-a106-60b3af7783ca-webhook-cert\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.302790 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65093a97-cc2b-4983-a106-60b3af7783ca-apiservice-cert\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.308329 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/65093a97-cc2b-4983-a106-60b3af7783ca-webhook-cert\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.309751 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/65093a97-cc2b-4983-a106-60b3af7783ca-apiservice-cert\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.348011 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdf6m\" (UniqueName: \"kubernetes.io/projected/65093a97-cc2b-4983-a106-60b3af7783ca-kube-api-access-hdf6m\") pod \"metallb-operator-controller-manager-75cb48855d-jh2hq\" (UID: \"65093a97-cc2b-4983-a106-60b3af7783ca\") " pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.489411 4829 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.511673 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.646794 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6884746998-tvx77"] Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.647839 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: W1002 07:29:23.650551 4829 reflector.go:561] object-"metallb-system"/"controller-dockercfg-n69dq": failed to list *v1.Secret: secrets "controller-dockercfg-n69dq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 07:29:23 crc kubenswrapper[4829]: E1002 07:29:23.650597 4829 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"controller-dockercfg-n69dq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"controller-dockercfg-n69dq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:29:23 crc kubenswrapper[4829]: W1002 07:29:23.650602 4829 reflector.go:561] object-"metallb-system"/"metallb-webhook-cert": failed to list *v1.Secret: secrets "metallb-webhook-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 07:29:23 crc kubenswrapper[4829]: E1002 07:29:23.650645 4829 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-webhook-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-webhook-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:29:23 crc kubenswrapper[4829]: W1002 07:29:23.652487 4829 reflector.go:561] object-"metallb-system"/"metallb-operator-webhook-server-service-cert": failed to list *v1.Secret: secrets "metallb-operator-webhook-server-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 07:29:23 crc kubenswrapper[4829]: E1002 07:29:23.652518 4829 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-webhook-server-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-webhook-server-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.670652 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6884746998-tvx77"] Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.809914 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twb6k\" (UniqueName: \"kubernetes.io/projected/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-kube-api-access-twb6k\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.809962 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-apiservice-cert\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.809982 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-webhook-cert\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.911574 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twb6k\" (UniqueName: \"kubernetes.io/projected/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-kube-api-access-twb6k\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.911627 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-apiservice-cert\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.911642 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-webhook-cert\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:23 crc kubenswrapper[4829]: I1002 07:29:23.931746 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twb6k\" (UniqueName: \"kubernetes.io/projected/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-kube-api-access-twb6k\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:24 crc kubenswrapper[4829]: I1002 07:29:24.096809 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq"] Oct 02 07:29:24 crc kubenswrapper[4829]: W1002 07:29:24.104382 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65093a97_cc2b_4983_a106_60b3af7783ca.slice/crio-70ebeb5229d971a97fed43f7dfb3ad03c92cff3d88eb4df7dcac400071978b20 WatchSource:0}: Error finding container 70ebeb5229d971a97fed43f7dfb3ad03c92cff3d88eb4df7dcac400071978b20: Status 404 returned error can't find the container with id 70ebeb5229d971a97fed43f7dfb3ad03c92cff3d88eb4df7dcac400071978b20 Oct 02 07:29:24 crc kubenswrapper[4829]: I1002 07:29:24.625465 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 07:29:24 crc kubenswrapper[4829]: I1002 07:29:24.903552 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 07:29:24 crc kubenswrapper[4829]: I1002 07:29:24.922925 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-apiservice-cert\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:24 crc kubenswrapper[4829]: I1002 07:29:24.926624 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/67cefc41-a1b6-4c77-9770-ea2c4f0fae83-webhook-cert\") pod \"metallb-operator-webhook-server-6884746998-tvx77\" (UID: \"67cefc41-a1b6-4c77-9770-ea2c4f0fae83\") " pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:25 crc kubenswrapper[4829]: I1002 07:29:25.010548 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-n69dq" Oct 02 07:29:25 crc kubenswrapper[4829]: I1002 07:29:25.011888 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" event={"ID":"65093a97-cc2b-4983-a106-60b3af7783ca","Type":"ContainerStarted","Data":"70ebeb5229d971a97fed43f7dfb3ad03c92cff3d88eb4df7dcac400071978b20"} Oct 02 07:29:25 crc kubenswrapper[4829]: I1002 07:29:25.184165 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:25 crc kubenswrapper[4829]: I1002 07:29:25.329377 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:29:25 crc kubenswrapper[4829]: I1002 07:29:25.329713 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:29:25 crc kubenswrapper[4829]: I1002 07:29:25.632355 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6884746998-tvx77"] Oct 02 07:29:26 crc kubenswrapper[4829]: I1002 07:29:26.018293 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" event={"ID":"67cefc41-a1b6-4c77-9770-ea2c4f0fae83","Type":"ContainerStarted","Data":"baf1bf9c4b4f5439982efda9ea3f9ba5c784dd91669334dc6fa6fb71fb3faa03"} Oct 02 07:29:28 crc kubenswrapper[4829]: I1002 07:29:28.043313 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" event={"ID":"65093a97-cc2b-4983-a106-60b3af7783ca","Type":"ContainerStarted","Data":"80a3156fb23f4fd01442398d8c5b068c7c71709a1d202f1046cb277ae62095c0"} Oct 02 07:29:28 crc kubenswrapper[4829]: I1002 07:29:28.053628 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:29:28 crc kubenswrapper[4829]: I1002 07:29:28.079751 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" podStartSLOduration=1.91060246 podStartE2EDuration="5.079730086s" podCreationTimestamp="2025-10-02 07:29:23 +0000 UTC" firstStartedPulling="2025-10-02 07:29:24.107587745 +0000 UTC m=+755.447236150" lastFinishedPulling="2025-10-02 07:29:27.276715371 +0000 UTC m=+758.616363776" observedRunningTime="2025-10-02 07:29:28.077795928 +0000 UTC m=+759.417444343" watchObservedRunningTime="2025-10-02 07:29:28.079730086 +0000 UTC m=+759.419378491" Oct 02 07:29:31 crc kubenswrapper[4829]: I1002 07:29:31.060298 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" event={"ID":"67cefc41-a1b6-4c77-9770-ea2c4f0fae83","Type":"ContainerStarted","Data":"c74ee67705b6f459c30b0c45b41427bf9aa6b259b679d2814dfb7fdb577ed24a"} Oct 02 07:29:31 crc kubenswrapper[4829]: I1002 07:29:31.060607 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:31 crc kubenswrapper[4829]: I1002 07:29:31.078623 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" podStartSLOduration=3.5399102989999998 podStartE2EDuration="8.078598392s" podCreationTimestamp="2025-10-02 07:29:23 +0000 UTC" firstStartedPulling="2025-10-02 07:29:25.646030159 +0000 UTC m=+756.985678564" lastFinishedPulling="2025-10-02 07:29:30.184718242 +0000 UTC m=+761.524366657" observedRunningTime="2025-10-02 07:29:31.075060807 +0000 UTC m=+762.414709242" watchObservedRunningTime="2025-10-02 07:29:31.078598392 +0000 UTC m=+762.418246817" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.391947 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7tn9w"] Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.394317 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.405677 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tn9w"] Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.497453 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-utilities\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.497520 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-catalog-content\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.497612 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clnkv\" (UniqueName: \"kubernetes.io/projected/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-kube-api-access-clnkv\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.599089 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-utilities\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.599452 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-catalog-content\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.599587 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clnkv\" (UniqueName: \"kubernetes.io/projected/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-kube-api-access-clnkv\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.599725 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-utilities\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.600379 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-catalog-content\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.627329 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clnkv\" (UniqueName: \"kubernetes.io/projected/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-kube-api-access-clnkv\") pod \"redhat-operators-7tn9w\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:37 crc kubenswrapper[4829]: I1002 07:29:37.713426 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:38 crc kubenswrapper[4829]: I1002 07:29:38.212167 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7tn9w"] Oct 02 07:29:38 crc kubenswrapper[4829]: W1002 07:29:38.228638 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76ff9c48_f3c2_4601_bbe0_a40efc6d682a.slice/crio-34bfce5cb7a225725445e4bc42d78d6c8bb7b0f7477407efdf2ae6d65e8f6968 WatchSource:0}: Error finding container 34bfce5cb7a225725445e4bc42d78d6c8bb7b0f7477407efdf2ae6d65e8f6968: Status 404 returned error can't find the container with id 34bfce5cb7a225725445e4bc42d78d6c8bb7b0f7477407efdf2ae6d65e8f6968 Oct 02 07:29:39 crc kubenswrapper[4829]: I1002 07:29:39.116640 4829 generic.go:334] "Generic (PLEG): container finished" podID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerID="9c7e7db318fbd57ea39c4ab31a2a3ebf4c28a8d87f749951fdf3d8c07796c64d" exitCode=0 Oct 02 07:29:39 crc kubenswrapper[4829]: I1002 07:29:39.117269 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerDied","Data":"9c7e7db318fbd57ea39c4ab31a2a3ebf4c28a8d87f749951fdf3d8c07796c64d"} Oct 02 07:29:39 crc kubenswrapper[4829]: I1002 07:29:39.117771 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerStarted","Data":"34bfce5cb7a225725445e4bc42d78d6c8bb7b0f7477407efdf2ae6d65e8f6968"} Oct 02 07:29:40 crc kubenswrapper[4829]: I1002 07:29:40.125702 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerStarted","Data":"d07b883d582313e623b51e0fb5134f2841e213a1e9d884be3b3c1d6218c01396"} Oct 02 07:29:41 crc kubenswrapper[4829]: I1002 07:29:41.132024 4829 generic.go:334] "Generic (PLEG): container finished" podID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerID="d07b883d582313e623b51e0fb5134f2841e213a1e9d884be3b3c1d6218c01396" exitCode=0 Oct 02 07:29:41 crc kubenswrapper[4829]: I1002 07:29:41.132074 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerDied","Data":"d07b883d582313e623b51e0fb5134f2841e213a1e9d884be3b3c1d6218c01396"} Oct 02 07:29:42 crc kubenswrapper[4829]: I1002 07:29:42.138827 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerStarted","Data":"9d17d52df576166dd68d973472a5ae0e36a426b754cf90c14a8a6fb32f233f0b"} Oct 02 07:29:42 crc kubenswrapper[4829]: I1002 07:29:42.167912 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7tn9w" podStartSLOduration=2.54491029 podStartE2EDuration="5.167896103s" podCreationTimestamp="2025-10-02 07:29:37 +0000 UTC" firstStartedPulling="2025-10-02 07:29:39.118654336 +0000 UTC m=+770.458302741" lastFinishedPulling="2025-10-02 07:29:41.741640139 +0000 UTC m=+773.081288554" observedRunningTime="2025-10-02 07:29:42.163577735 +0000 UTC m=+773.503226160" watchObservedRunningTime="2025-10-02 07:29:42.167896103 +0000 UTC m=+773.507544508" Oct 02 07:29:45 crc kubenswrapper[4829]: I1002 07:29:45.193757 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6884746998-tvx77" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.524924 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pv5qf"] Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.526409 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.541310 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pv5qf"] Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.627568 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866eefa0-245f-4ed1-b93c-000cf03733b0-catalog-content\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.627619 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk8sf\" (UniqueName: \"kubernetes.io/projected/866eefa0-245f-4ed1-b93c-000cf03733b0-kube-api-access-qk8sf\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.627686 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866eefa0-245f-4ed1-b93c-000cf03733b0-utilities\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.714284 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.714443 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.729093 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866eefa0-245f-4ed1-b93c-000cf03733b0-catalog-content\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.729152 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk8sf\" (UniqueName: \"kubernetes.io/projected/866eefa0-245f-4ed1-b93c-000cf03733b0-kube-api-access-qk8sf\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.729211 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866eefa0-245f-4ed1-b93c-000cf03733b0-utilities\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.729738 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/866eefa0-245f-4ed1-b93c-000cf03733b0-utilities\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.729790 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/866eefa0-245f-4ed1-b93c-000cf03733b0-catalog-content\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.760654 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.761193 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk8sf\" (UniqueName: \"kubernetes.io/projected/866eefa0-245f-4ed1-b93c-000cf03733b0-kube-api-access-qk8sf\") pod \"certified-operators-pv5qf\" (UID: \"866eefa0-245f-4ed1-b93c-000cf03733b0\") " pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:47 crc kubenswrapper[4829]: I1002 07:29:47.844209 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:48 crc kubenswrapper[4829]: I1002 07:29:48.218127 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:48 crc kubenswrapper[4829]: I1002 07:29:48.309025 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pv5qf"] Oct 02 07:29:49 crc kubenswrapper[4829]: I1002 07:29:49.188222 4829 generic.go:334] "Generic (PLEG): container finished" podID="866eefa0-245f-4ed1-b93c-000cf03733b0" containerID="a0011d834b50902cf781f9c38055f3aff4565814d531031b45d58807df632b31" exitCode=0 Oct 02 07:29:49 crc kubenswrapper[4829]: I1002 07:29:49.188319 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5qf" event={"ID":"866eefa0-245f-4ed1-b93c-000cf03733b0","Type":"ContainerDied","Data":"a0011d834b50902cf781f9c38055f3aff4565814d531031b45d58807df632b31"} Oct 02 07:29:49 crc kubenswrapper[4829]: I1002 07:29:49.188636 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5qf" event={"ID":"866eefa0-245f-4ed1-b93c-000cf03733b0","Type":"ContainerStarted","Data":"3bac00a142256ed0c445bcb82fe2aa3eb294de3ba592970a4819ba9c8e2a81dd"} Oct 02 07:29:50 crc kubenswrapper[4829]: I1002 07:29:50.905570 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tn9w"] Oct 02 07:29:51 crc kubenswrapper[4829]: I1002 07:29:51.201093 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7tn9w" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="registry-server" containerID="cri-o://9d17d52df576166dd68d973472a5ae0e36a426b754cf90c14a8a6fb32f233f0b" gracePeriod=2 Oct 02 07:29:52 crc kubenswrapper[4829]: I1002 07:29:52.208387 4829 generic.go:334] "Generic (PLEG): container finished" podID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerID="9d17d52df576166dd68d973472a5ae0e36a426b754cf90c14a8a6fb32f233f0b" exitCode=0 Oct 02 07:29:52 crc kubenswrapper[4829]: I1002 07:29:52.208451 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerDied","Data":"9d17d52df576166dd68d973472a5ae0e36a426b754cf90c14a8a6fb32f233f0b"} Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.161422 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.217991 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7tn9w" event={"ID":"76ff9c48-f3c2-4601-bbe0-a40efc6d682a","Type":"ContainerDied","Data":"34bfce5cb7a225725445e4bc42d78d6c8bb7b0f7477407efdf2ae6d65e8f6968"} Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.218053 4829 scope.go:117] "RemoveContainer" containerID="9d17d52df576166dd68d973472a5ae0e36a426b754cf90c14a8a6fb32f233f0b" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.218081 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7tn9w" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.222044 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5qf" event={"ID":"866eefa0-245f-4ed1-b93c-000cf03733b0","Type":"ContainerStarted","Data":"64984757b4b7adc454f397d404bcec4b1fb2895ceefd274759ef2cd8b38e9764"} Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.244347 4829 scope.go:117] "RemoveContainer" containerID="d07b883d582313e623b51e0fb5134f2841e213a1e9d884be3b3c1d6218c01396" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.262907 4829 scope.go:117] "RemoveContainer" containerID="9c7e7db318fbd57ea39c4ab31a2a3ebf4c28a8d87f749951fdf3d8c07796c64d" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.304416 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-catalog-content\") pod \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.304519 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-utilities\") pod \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.304550 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clnkv\" (UniqueName: \"kubernetes.io/projected/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-kube-api-access-clnkv\") pod \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\" (UID: \"76ff9c48-f3c2-4601-bbe0-a40efc6d682a\") " Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.307025 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-utilities" (OuterVolumeSpecName: "utilities") pod "76ff9c48-f3c2-4601-bbe0-a40efc6d682a" (UID: "76ff9c48-f3c2-4601-bbe0-a40efc6d682a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.314741 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-kube-api-access-clnkv" (OuterVolumeSpecName: "kube-api-access-clnkv") pod "76ff9c48-f3c2-4601-bbe0-a40efc6d682a" (UID: "76ff9c48-f3c2-4601-bbe0-a40efc6d682a"). InnerVolumeSpecName "kube-api-access-clnkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.406906 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.406960 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clnkv\" (UniqueName: \"kubernetes.io/projected/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-kube-api-access-clnkv\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.409136 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76ff9c48-f3c2-4601-bbe0-a40efc6d682a" (UID: "76ff9c48-f3c2-4601-bbe0-a40efc6d682a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.508896 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ff9c48-f3c2-4601-bbe0-a40efc6d682a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.548497 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7tn9w"] Oct 02 07:29:53 crc kubenswrapper[4829]: I1002 07:29:53.553113 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7tn9w"] Oct 02 07:29:54 crc kubenswrapper[4829]: I1002 07:29:54.231309 4829 generic.go:334] "Generic (PLEG): container finished" podID="866eefa0-245f-4ed1-b93c-000cf03733b0" containerID="64984757b4b7adc454f397d404bcec4b1fb2895ceefd274759ef2cd8b38e9764" exitCode=0 Oct 02 07:29:54 crc kubenswrapper[4829]: I1002 07:29:54.231352 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5qf" event={"ID":"866eefa0-245f-4ed1-b93c-000cf03733b0","Type":"ContainerDied","Data":"64984757b4b7adc454f397d404bcec4b1fb2895ceefd274759ef2cd8b38e9764"} Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.247164 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pv5qf" event={"ID":"866eefa0-245f-4ed1-b93c-000cf03733b0","Type":"ContainerStarted","Data":"ca9bf01fb1b09ead64a7d1a2f98eb4f9ac0a8dde630e11948dde2114c1035d53"} Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.273605 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pv5qf" podStartSLOduration=2.812187738 podStartE2EDuration="8.273585163s" podCreationTimestamp="2025-10-02 07:29:47 +0000 UTC" firstStartedPulling="2025-10-02 07:29:49.190890537 +0000 UTC m=+780.530538972" lastFinishedPulling="2025-10-02 07:29:54.652287982 +0000 UTC m=+785.991936397" observedRunningTime="2025-10-02 07:29:55.272691674 +0000 UTC m=+786.612340089" watchObservedRunningTime="2025-10-02 07:29:55.273585163 +0000 UTC m=+786.613233578" Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.329120 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.329177 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.329241 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.329847 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d8e666d4540458f8abd1af2cd0bebf671777110d93b151c9aedce3fa191263d9"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.329930 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://d8e666d4540458f8abd1af2cd0bebf671777110d93b151c9aedce3fa191263d9" gracePeriod=600 Oct 02 07:29:55 crc kubenswrapper[4829]: I1002 07:29:55.473929 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" path="/var/lib/kubelet/pods/76ff9c48-f3c2-4601-bbe0-a40efc6d682a/volumes" Oct 02 07:29:56 crc kubenswrapper[4829]: I1002 07:29:56.254220 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="d8e666d4540458f8abd1af2cd0bebf671777110d93b151c9aedce3fa191263d9" exitCode=0 Oct 02 07:29:56 crc kubenswrapper[4829]: I1002 07:29:56.255080 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"d8e666d4540458f8abd1af2cd0bebf671777110d93b151c9aedce3fa191263d9"} Oct 02 07:29:56 crc kubenswrapper[4829]: I1002 07:29:56.255103 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"34b90026201c24a6d0bc1440495bb6980e11f5b356c27ed499750d010a61ec35"} Oct 02 07:29:56 crc kubenswrapper[4829]: I1002 07:29:56.255120 4829 scope.go:117] "RemoveContainer" containerID="3c344e588de3b2b45a68b8eaf305327aebb97a91305e92834159bf1c925efb2c" Oct 02 07:29:57 crc kubenswrapper[4829]: I1002 07:29:57.845039 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:57 crc kubenswrapper[4829]: I1002 07:29:57.845432 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:29:57 crc kubenswrapper[4829]: I1002 07:29:57.907774 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.169446 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf"] Oct 02 07:30:00 crc kubenswrapper[4829]: E1002 07:30:00.170290 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="registry-server" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.170309 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="registry-server" Oct 02 07:30:00 crc kubenswrapper[4829]: E1002 07:30:00.170354 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="extract-utilities" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.170364 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="extract-utilities" Oct 02 07:30:00 crc kubenswrapper[4829]: E1002 07:30:00.170390 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="extract-content" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.170399 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="extract-content" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.170695 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ff9c48-f3c2-4601-bbe0-a40efc6d682a" containerName="registry-server" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.171484 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.175731 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.176242 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.192217 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf"] Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.320576 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c3f20db-ce94-41ff-835f-86301d1ef391-config-volume\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.320637 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c3f20db-ce94-41ff-835f-86301d1ef391-secret-volume\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.320678 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbn5h\" (UniqueName: \"kubernetes.io/projected/7c3f20db-ce94-41ff-835f-86301d1ef391-kube-api-access-kbn5h\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.422390 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbn5h\" (UniqueName: \"kubernetes.io/projected/7c3f20db-ce94-41ff-835f-86301d1ef391-kube-api-access-kbn5h\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.422499 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c3f20db-ce94-41ff-835f-86301d1ef391-config-volume\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.422542 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c3f20db-ce94-41ff-835f-86301d1ef391-secret-volume\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.423663 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c3f20db-ce94-41ff-835f-86301d1ef391-config-volume\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.429149 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c3f20db-ce94-41ff-835f-86301d1ef391-secret-volume\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.438888 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbn5h\" (UniqueName: \"kubernetes.io/projected/7c3f20db-ce94-41ff-835f-86301d1ef391-kube-api-access-kbn5h\") pod \"collect-profiles-29323170-2tzgf\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.503365 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:00 crc kubenswrapper[4829]: I1002 07:30:00.925711 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf"] Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.291801 4829 generic.go:334] "Generic (PLEG): container finished" podID="7c3f20db-ce94-41ff-835f-86301d1ef391" containerID="927babadc432a63f7a214d1234b54b592230d6d84e17c703028640b6a553aeba" exitCode=0 Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.291956 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" event={"ID":"7c3f20db-ce94-41ff-835f-86301d1ef391","Type":"ContainerDied","Data":"927babadc432a63f7a214d1234b54b592230d6d84e17c703028640b6a553aeba"} Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.292089 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" event={"ID":"7c3f20db-ce94-41ff-835f-86301d1ef391","Type":"ContainerStarted","Data":"67e35bcd3def7e7eb2ff48bc22ce651becabb426b0a253a89fcd49be019a75ff"} Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.425621 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7qv75"] Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.426883 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.440991 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qv75"] Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.536687 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-utilities\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.536872 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-catalog-content\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.536905 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxb8b\" (UniqueName: \"kubernetes.io/projected/e7b475c1-50d5-4956-b9f4-dd90f360da8f-kube-api-access-qxb8b\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.638900 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-catalog-content\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.638978 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxb8b\" (UniqueName: \"kubernetes.io/projected/e7b475c1-50d5-4956-b9f4-dd90f360da8f-kube-api-access-qxb8b\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.639029 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-utilities\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.639762 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-catalog-content\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.639824 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-utilities\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.665539 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxb8b\" (UniqueName: \"kubernetes.io/projected/e7b475c1-50d5-4956-b9f4-dd90f360da8f-kube-api-access-qxb8b\") pod \"redhat-marketplace-7qv75\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:01 crc kubenswrapper[4829]: I1002 07:30:01.791357 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.190097 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qv75"] Oct 02 07:30:02 crc kubenswrapper[4829]: W1002 07:30:02.196448 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7b475c1_50d5_4956_b9f4_dd90f360da8f.slice/crio-9d8a6004e64f990ca5da1fb3e9689b458621579a59fa3459a15d867563b54816 WatchSource:0}: Error finding container 9d8a6004e64f990ca5da1fb3e9689b458621579a59fa3459a15d867563b54816: Status 404 returned error can't find the container with id 9d8a6004e64f990ca5da1fb3e9689b458621579a59fa3459a15d867563b54816 Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.299142 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qv75" event={"ID":"e7b475c1-50d5-4956-b9f4-dd90f360da8f","Type":"ContainerStarted","Data":"9d8a6004e64f990ca5da1fb3e9689b458621579a59fa3459a15d867563b54816"} Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.541125 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.651584 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c3f20db-ce94-41ff-835f-86301d1ef391-secret-volume\") pod \"7c3f20db-ce94-41ff-835f-86301d1ef391\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.651677 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbn5h\" (UniqueName: \"kubernetes.io/projected/7c3f20db-ce94-41ff-835f-86301d1ef391-kube-api-access-kbn5h\") pod \"7c3f20db-ce94-41ff-835f-86301d1ef391\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.651733 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c3f20db-ce94-41ff-835f-86301d1ef391-config-volume\") pod \"7c3f20db-ce94-41ff-835f-86301d1ef391\" (UID: \"7c3f20db-ce94-41ff-835f-86301d1ef391\") " Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.652623 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3f20db-ce94-41ff-835f-86301d1ef391-config-volume" (OuterVolumeSpecName: "config-volume") pod "7c3f20db-ce94-41ff-835f-86301d1ef391" (UID: "7c3f20db-ce94-41ff-835f-86301d1ef391"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.656962 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c3f20db-ce94-41ff-835f-86301d1ef391-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7c3f20db-ce94-41ff-835f-86301d1ef391" (UID: "7c3f20db-ce94-41ff-835f-86301d1ef391"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.657412 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3f20db-ce94-41ff-835f-86301d1ef391-kube-api-access-kbn5h" (OuterVolumeSpecName: "kube-api-access-kbn5h") pod "7c3f20db-ce94-41ff-835f-86301d1ef391" (UID: "7c3f20db-ce94-41ff-835f-86301d1ef391"). InnerVolumeSpecName "kube-api-access-kbn5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.753598 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbn5h\" (UniqueName: \"kubernetes.io/projected/7c3f20db-ce94-41ff-835f-86301d1ef391-kube-api-access-kbn5h\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.753639 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7c3f20db-ce94-41ff-835f-86301d1ef391-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:02 crc kubenswrapper[4829]: I1002 07:30:02.753650 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7c3f20db-ce94-41ff-835f-86301d1ef391-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:03 crc kubenswrapper[4829]: I1002 07:30:03.306249 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" Oct 02 07:30:03 crc kubenswrapper[4829]: I1002 07:30:03.306220 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf" event={"ID":"7c3f20db-ce94-41ff-835f-86301d1ef391","Type":"ContainerDied","Data":"67e35bcd3def7e7eb2ff48bc22ce651becabb426b0a253a89fcd49be019a75ff"} Oct 02 07:30:03 crc kubenswrapper[4829]: I1002 07:30:03.306424 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67e35bcd3def7e7eb2ff48bc22ce651becabb426b0a253a89fcd49be019a75ff" Oct 02 07:30:03 crc kubenswrapper[4829]: I1002 07:30:03.307920 4829 generic.go:334] "Generic (PLEG): container finished" podID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerID="613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613" exitCode=0 Oct 02 07:30:03 crc kubenswrapper[4829]: I1002 07:30:03.307957 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qv75" event={"ID":"e7b475c1-50d5-4956-b9f4-dd90f360da8f","Type":"ContainerDied","Data":"613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613"} Oct 02 07:30:03 crc kubenswrapper[4829]: I1002 07:30:03.515259 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-75cb48855d-jh2hq" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.285423 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-8sxd9"] Oct 02 07:30:04 crc kubenswrapper[4829]: E1002 07:30:04.286020 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3f20db-ce94-41ff-835f-86301d1ef391" containerName="collect-profiles" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.286044 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3f20db-ce94-41ff-835f-86301d1ef391" containerName="collect-profiles" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.286252 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3f20db-ce94-41ff-835f-86301d1ef391" containerName="collect-profiles" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.292954 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.293975 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2"] Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.298536 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.299605 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.299779 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.299930 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-gpvfs" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.305151 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2"] Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.305449 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372012 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-reloader\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372060 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99893ca5-1605-4a55-9ce8-6711e86fa31b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t78d2\" (UID: \"99893ca5-1605-4a55-9ce8-6711e86fa31b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372083 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-sockets\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372117 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a85bf9a0-7c63-492e-8662-c463be70aad0-metrics-certs\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372133 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j59wp\" (UniqueName: \"kubernetes.io/projected/a85bf9a0-7c63-492e-8662-c463be70aad0-kube-api-access-j59wp\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372168 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgpnf\" (UniqueName: \"kubernetes.io/projected/99893ca5-1605-4a55-9ce8-6711e86fa31b-kube-api-access-fgpnf\") pod \"frr-k8s-webhook-server-64bf5d555-t78d2\" (UID: \"99893ca5-1605-4a55-9ce8-6711e86fa31b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372182 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-conf\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372204 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-startup\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.372243 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-metrics\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.384294 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-mvlxd"] Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.385527 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.388279 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.388534 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-p4mx2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.388561 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.389084 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.401711 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-mpn22"] Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.402764 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.406309 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.428011 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-mpn22"] Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473689 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgbr4\" (UniqueName: \"kubernetes.io/projected/321a9e53-9458-4d2f-9784-15f96c7d157b-kube-api-access-wgbr4\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473743 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a85bf9a0-7c63-492e-8662-c463be70aad0-metrics-certs\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473765 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j59wp\" (UniqueName: \"kubernetes.io/projected/a85bf9a0-7c63-492e-8662-c463be70aad0-kube-api-access-j59wp\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473783 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbv62\" (UniqueName: \"kubernetes.io/projected/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-kube-api-access-mbv62\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473817 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgpnf\" (UniqueName: \"kubernetes.io/projected/99893ca5-1605-4a55-9ce8-6711e86fa31b-kube-api-access-fgpnf\") pod \"frr-k8s-webhook-server-64bf5d555-t78d2\" (UID: \"99893ca5-1605-4a55-9ce8-6711e86fa31b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473831 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-conf\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473845 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-metrics-certs\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473865 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/321a9e53-9458-4d2f-9784-15f96c7d157b-cert\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473884 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473899 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-startup\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473923 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-metrics\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473944 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-reloader\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473966 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99893ca5-1605-4a55-9ce8-6711e86fa31b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t78d2\" (UID: \"99893ca5-1605-4a55-9ce8-6711e86fa31b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.473986 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/321a9e53-9458-4d2f-9784-15f96c7d157b-metrics-certs\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.474001 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-sockets\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.474018 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-metallb-excludel2\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.474798 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-conf\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.475013 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-metrics\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.475132 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-reloader\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.475494 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-sockets\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.475694 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/a85bf9a0-7c63-492e-8662-c463be70aad0-frr-startup\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.490586 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/99893ca5-1605-4a55-9ce8-6711e86fa31b-cert\") pod \"frr-k8s-webhook-server-64bf5d555-t78d2\" (UID: \"99893ca5-1605-4a55-9ce8-6711e86fa31b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.493354 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j59wp\" (UniqueName: \"kubernetes.io/projected/a85bf9a0-7c63-492e-8662-c463be70aad0-kube-api-access-j59wp\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.493561 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgpnf\" (UniqueName: \"kubernetes.io/projected/99893ca5-1605-4a55-9ce8-6711e86fa31b-kube-api-access-fgpnf\") pod \"frr-k8s-webhook-server-64bf5d555-t78d2\" (UID: \"99893ca5-1605-4a55-9ce8-6711e86fa31b\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.501661 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a85bf9a0-7c63-492e-8662-c463be70aad0-metrics-certs\") pod \"frr-k8s-8sxd9\" (UID: \"a85bf9a0-7c63-492e-8662-c463be70aad0\") " pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574739 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/321a9e53-9458-4d2f-9784-15f96c7d157b-metrics-certs\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574792 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-metallb-excludel2\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574830 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgbr4\" (UniqueName: \"kubernetes.io/projected/321a9e53-9458-4d2f-9784-15f96c7d157b-kube-api-access-wgbr4\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574850 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbv62\" (UniqueName: \"kubernetes.io/projected/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-kube-api-access-mbv62\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574884 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-metrics-certs\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574904 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/321a9e53-9458-4d2f-9784-15f96c7d157b-cert\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.574925 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: E1002 07:30:04.575013 4829 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 07:30:04 crc kubenswrapper[4829]: E1002 07:30:04.575061 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist podName:3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64 nodeName:}" failed. No retries permitted until 2025-10-02 07:30:05.075046057 +0000 UTC m=+796.414694462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist") pod "speaker-mvlxd" (UID: "3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64") : secret "metallb-memberlist" not found Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.575771 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-metallb-excludel2\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.577833 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/321a9e53-9458-4d2f-9784-15f96c7d157b-metrics-certs\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.578000 4829 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.579696 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-metrics-certs\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.590820 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/321a9e53-9458-4d2f-9784-15f96c7d157b-cert\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.591156 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbv62\" (UniqueName: \"kubernetes.io/projected/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-kube-api-access-mbv62\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.593263 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgbr4\" (UniqueName: \"kubernetes.io/projected/321a9e53-9458-4d2f-9784-15f96c7d157b-kube-api-access-wgbr4\") pod \"controller-68d546b9d8-mpn22\" (UID: \"321a9e53-9458-4d2f-9784-15f96c7d157b\") " pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.667505 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.685715 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.726856 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:04 crc kubenswrapper[4829]: I1002 07:30:04.943774 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2"] Oct 02 07:30:04 crc kubenswrapper[4829]: W1002 07:30:04.957616 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99893ca5_1605_4a55_9ce8_6711e86fa31b.slice/crio-b9afcdfc88937db1177312f350e653a6d2b06576b4242a997ace57ddd993275a WatchSource:0}: Error finding container b9afcdfc88937db1177312f350e653a6d2b06576b4242a997ace57ddd993275a: Status 404 returned error can't find the container with id b9afcdfc88937db1177312f350e653a6d2b06576b4242a997ace57ddd993275a Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.083671 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:05 crc kubenswrapper[4829]: E1002 07:30:05.083873 4829 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 07:30:05 crc kubenswrapper[4829]: E1002 07:30:05.083949 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist podName:3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64 nodeName:}" failed. No retries permitted until 2025-10-02 07:30:06.083928926 +0000 UTC m=+797.423577331 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist") pod "speaker-mvlxd" (UID: "3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64") : secret "metallb-memberlist" not found Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.238543 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-mpn22"] Oct 02 07:30:05 crc kubenswrapper[4829]: W1002 07:30:05.245756 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod321a9e53_9458_4d2f_9784_15f96c7d157b.slice/crio-dd50089a1d8ec785b0a288e678a98bc9fac869025b6baa54efde7fc643a8891f WatchSource:0}: Error finding container dd50089a1d8ec785b0a288e678a98bc9fac869025b6baa54efde7fc643a8891f: Status 404 returned error can't find the container with id dd50089a1d8ec785b0a288e678a98bc9fac869025b6baa54efde7fc643a8891f Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.341380 4829 generic.go:334] "Generic (PLEG): container finished" podID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerID="4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1" exitCode=0 Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.341513 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qv75" event={"ID":"e7b475c1-50d5-4956-b9f4-dd90f360da8f","Type":"ContainerDied","Data":"4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1"} Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.343634 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-mpn22" event={"ID":"321a9e53-9458-4d2f-9784-15f96c7d157b","Type":"ContainerStarted","Data":"dd50089a1d8ec785b0a288e678a98bc9fac869025b6baa54efde7fc643a8891f"} Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.346041 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" event={"ID":"99893ca5-1605-4a55-9ce8-6711e86fa31b","Type":"ContainerStarted","Data":"b9afcdfc88937db1177312f350e653a6d2b06576b4242a997ace57ddd993275a"} Oct 02 07:30:05 crc kubenswrapper[4829]: I1002 07:30:05.347580 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"0f5f2a0bb50b97b971fd4b1b26c96e17272efbc2859e1a676678d1450d9dfa05"} Oct 02 07:30:06 crc kubenswrapper[4829]: I1002 07:30:06.099798 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:06 crc kubenswrapper[4829]: E1002 07:30:06.099987 4829 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 07:30:06 crc kubenswrapper[4829]: E1002 07:30:06.100051 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist podName:3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64 nodeName:}" failed. No retries permitted until 2025-10-02 07:30:08.10003522 +0000 UTC m=+799.439683635 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist") pod "speaker-mvlxd" (UID: "3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64") : secret "metallb-memberlist" not found Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.369176 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-mpn22" event={"ID":"321a9e53-9458-4d2f-9784-15f96c7d157b","Type":"ContainerStarted","Data":"ea51650e615661c5c316c456fd3bccc64127734238869e65cbbfbefb74ce9fa2"} Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.369236 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-mpn22" event={"ID":"321a9e53-9458-4d2f-9784-15f96c7d157b","Type":"ContainerStarted","Data":"4eb9dcdf6e374a689ec34987600859cc254f6b0440037c9497d1096758991cce"} Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.369394 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.372391 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qv75" event={"ID":"e7b475c1-50d5-4956-b9f4-dd90f360da8f","Type":"ContainerStarted","Data":"28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105"} Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.420602 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-mpn22" podStartSLOduration=3.420574855 podStartE2EDuration="3.420574855s" podCreationTimestamp="2025-10-02 07:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:30:07.400641054 +0000 UTC m=+798.740289459" watchObservedRunningTime="2025-10-02 07:30:07.420574855 +0000 UTC m=+798.760223260" Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.426337 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7qv75" podStartSLOduration=2.972860173 podStartE2EDuration="6.42632653s" podCreationTimestamp="2025-10-02 07:30:01 +0000 UTC" firstStartedPulling="2025-10-02 07:30:03.309745118 +0000 UTC m=+794.649393533" lastFinishedPulling="2025-10-02 07:30:06.763211485 +0000 UTC m=+798.102859890" observedRunningTime="2025-10-02 07:30:07.421199385 +0000 UTC m=+798.760847790" watchObservedRunningTime="2025-10-02 07:30:07.42632653 +0000 UTC m=+798.765974935" Oct 02 07:30:07 crc kubenswrapper[4829]: I1002 07:30:07.912073 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pv5qf" Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.137889 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.148532 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64-memberlist\") pod \"speaker-mvlxd\" (UID: \"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64\") " pod="metallb-system/speaker-mvlxd" Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.313602 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-mvlxd" Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.380597 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mvlxd" event={"ID":"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64","Type":"ContainerStarted","Data":"c28c115e24d3366a890c473710db3555bb514247d9ca35c0123deeb5d45f4f3b"} Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.634226 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pv5qf"] Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.822124 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfj6n"] Oct 02 07:30:08 crc kubenswrapper[4829]: I1002 07:30:08.822355 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kfj6n" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="registry-server" containerID="cri-o://94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698" gracePeriod=2 Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.291105 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.388305 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mvlxd" event={"ID":"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64","Type":"ContainerStarted","Data":"006a2a45bcfd3552584c077adde72d0d77eacbfd3418b8afb3d39012e5c46901"} Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.388365 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-mvlxd" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.388381 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-mvlxd" event={"ID":"3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64","Type":"ContainerStarted","Data":"8af289a9d0bfd097e01f51a584147cc45ba180ff203c43f89483b85a68693904"} Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.391822 4829 generic.go:334] "Generic (PLEG): container finished" podID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerID="94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698" exitCode=0 Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.391864 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfj6n" event={"ID":"f8d76d5c-3014-461e-8bac-253a85703dd9","Type":"ContainerDied","Data":"94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698"} Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.391896 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kfj6n" event={"ID":"f8d76d5c-3014-461e-8bac-253a85703dd9","Type":"ContainerDied","Data":"2e8d028765e7c4b21628d56ad9314d70cce671f1a165644accc2e603290b5ead"} Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.391917 4829 scope.go:117] "RemoveContainer" containerID="94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.392032 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kfj6n" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.405723 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-mvlxd" podStartSLOduration=5.40570903 podStartE2EDuration="5.40570903s" podCreationTimestamp="2025-10-02 07:30:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:30:09.404528812 +0000 UTC m=+800.744177217" watchObservedRunningTime="2025-10-02 07:30:09.40570903 +0000 UTC m=+800.745357435" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.418621 4829 scope.go:117] "RemoveContainer" containerID="1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.443993 4829 scope.go:117] "RemoveContainer" containerID="c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.457638 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-catalog-content\") pod \"f8d76d5c-3014-461e-8bac-253a85703dd9\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.457756 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cjl7\" (UniqueName: \"kubernetes.io/projected/f8d76d5c-3014-461e-8bac-253a85703dd9-kube-api-access-2cjl7\") pod \"f8d76d5c-3014-461e-8bac-253a85703dd9\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.457834 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-utilities\") pod \"f8d76d5c-3014-461e-8bac-253a85703dd9\" (UID: \"f8d76d5c-3014-461e-8bac-253a85703dd9\") " Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.459802 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-utilities" (OuterVolumeSpecName: "utilities") pod "f8d76d5c-3014-461e-8bac-253a85703dd9" (UID: "f8d76d5c-3014-461e-8bac-253a85703dd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.463976 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8d76d5c-3014-461e-8bac-253a85703dd9-kube-api-access-2cjl7" (OuterVolumeSpecName: "kube-api-access-2cjl7") pod "f8d76d5c-3014-461e-8bac-253a85703dd9" (UID: "f8d76d5c-3014-461e-8bac-253a85703dd9"). InnerVolumeSpecName "kube-api-access-2cjl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.493390 4829 scope.go:117] "RemoveContainer" containerID="94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698" Oct 02 07:30:09 crc kubenswrapper[4829]: E1002 07:30:09.493746 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698\": container with ID starting with 94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698 not found: ID does not exist" containerID="94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.493783 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698"} err="failed to get container status \"94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698\": rpc error: code = NotFound desc = could not find container \"94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698\": container with ID starting with 94044ca46a0072739d30d489cde97a78c85bb7f14c2f42f3a5228ad2db6ae698 not found: ID does not exist" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.493804 4829 scope.go:117] "RemoveContainer" containerID="1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380" Oct 02 07:30:09 crc kubenswrapper[4829]: E1002 07:30:09.494029 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380\": container with ID starting with 1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380 not found: ID does not exist" containerID="1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.494064 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380"} err="failed to get container status \"1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380\": rpc error: code = NotFound desc = could not find container \"1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380\": container with ID starting with 1072cfc1b68e56c4ab09fc0a35f9a0c370a6e6ecbeebbfe10e953ec5d5324380 not found: ID does not exist" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.494082 4829 scope.go:117] "RemoveContainer" containerID="c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132" Oct 02 07:30:09 crc kubenswrapper[4829]: E1002 07:30:09.494288 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132\": container with ID starting with c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132 not found: ID does not exist" containerID="c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.494311 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132"} err="failed to get container status \"c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132\": rpc error: code = NotFound desc = could not find container \"c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132\": container with ID starting with c89368dd61903045148f5b4a239f2bcb4237c1402c3d0ae0f9b1fcd00761b132 not found: ID does not exist" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.518567 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8d76d5c-3014-461e-8bac-253a85703dd9" (UID: "f8d76d5c-3014-461e-8bac-253a85703dd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.559389 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.559538 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8d76d5c-3014-461e-8bac-253a85703dd9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.559683 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cjl7\" (UniqueName: \"kubernetes.io/projected/f8d76d5c-3014-461e-8bac-253a85703dd9-kube-api-access-2cjl7\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.725021 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kfj6n"] Oct 02 07:30:09 crc kubenswrapper[4829]: I1002 07:30:09.730861 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kfj6n"] Oct 02 07:30:11 crc kubenswrapper[4829]: I1002 07:30:11.469941 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" path="/var/lib/kubelet/pods/f8d76d5c-3014-461e-8bac-253a85703dd9/volumes" Oct 02 07:30:11 crc kubenswrapper[4829]: I1002 07:30:11.791729 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:11 crc kubenswrapper[4829]: I1002 07:30:11.791772 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:11 crc kubenswrapper[4829]: I1002 07:30:11.860129 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:12 crc kubenswrapper[4829]: I1002 07:30:12.474198 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.003400 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qv75"] Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.442255 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" event={"ID":"99893ca5-1605-4a55-9ce8-6711e86fa31b","Type":"ContainerStarted","Data":"77b4dede69c7faf8d438e764572212ffe4b87e344a1a6632dc3c48a5d500af5f"} Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.442432 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.443865 4829 generic.go:334] "Generic (PLEG): container finished" podID="a85bf9a0-7c63-492e-8662-c463be70aad0" containerID="92d49460873c4907acd1314eee6fc9ea43c6204d7d6972fe1314119d2d375cf1" exitCode=0 Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.443918 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerDied","Data":"92d49460873c4907acd1314eee6fc9ea43c6204d7d6972fe1314119d2d375cf1"} Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.444086 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7qv75" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="registry-server" containerID="cri-o://28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105" gracePeriod=2 Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.460569 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" podStartSLOduration=2.013784638 podStartE2EDuration="10.460551884s" podCreationTimestamp="2025-10-02 07:30:04 +0000 UTC" firstStartedPulling="2025-10-02 07:30:04.960930674 +0000 UTC m=+796.300579079" lastFinishedPulling="2025-10-02 07:30:13.40769791 +0000 UTC m=+804.747346325" observedRunningTime="2025-10-02 07:30:14.455630937 +0000 UTC m=+805.795279352" watchObservedRunningTime="2025-10-02 07:30:14.460551884 +0000 UTC m=+805.800200299" Oct 02 07:30:14 crc kubenswrapper[4829]: I1002 07:30:14.869918 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.052964 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-catalog-content\") pod \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.053094 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-utilities\") pod \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.053205 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxb8b\" (UniqueName: \"kubernetes.io/projected/e7b475c1-50d5-4956-b9f4-dd90f360da8f-kube-api-access-qxb8b\") pod \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\" (UID: \"e7b475c1-50d5-4956-b9f4-dd90f360da8f\") " Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.054907 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-utilities" (OuterVolumeSpecName: "utilities") pod "e7b475c1-50d5-4956-b9f4-dd90f360da8f" (UID: "e7b475c1-50d5-4956-b9f4-dd90f360da8f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.061450 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b475c1-50d5-4956-b9f4-dd90f360da8f-kube-api-access-qxb8b" (OuterVolumeSpecName: "kube-api-access-qxb8b") pod "e7b475c1-50d5-4956-b9f4-dd90f360da8f" (UID: "e7b475c1-50d5-4956-b9f4-dd90f360da8f"). InnerVolumeSpecName "kube-api-access-qxb8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.066336 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7b475c1-50d5-4956-b9f4-dd90f360da8f" (UID: "e7b475c1-50d5-4956-b9f4-dd90f360da8f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.155467 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxb8b\" (UniqueName: \"kubernetes.io/projected/e7b475c1-50d5-4956-b9f4-dd90f360da8f-kube-api-access-qxb8b\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.155509 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.155519 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7b475c1-50d5-4956-b9f4-dd90f360da8f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.454364 4829 generic.go:334] "Generic (PLEG): container finished" podID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerID="28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105" exitCode=0 Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.454506 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7qv75" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.454542 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qv75" event={"ID":"e7b475c1-50d5-4956-b9f4-dd90f360da8f","Type":"ContainerDied","Data":"28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105"} Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.455294 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7qv75" event={"ID":"e7b475c1-50d5-4956-b9f4-dd90f360da8f","Type":"ContainerDied","Data":"9d8a6004e64f990ca5da1fb3e9689b458621579a59fa3459a15d867563b54816"} Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.455336 4829 scope.go:117] "RemoveContainer" containerID="28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.463503 4829 generic.go:334] "Generic (PLEG): container finished" podID="a85bf9a0-7c63-492e-8662-c463be70aad0" containerID="2d0182fb84b8841913b5e519c4275a805d1ab67d27a93b23760546b6f43b887b" exitCode=0 Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.478434 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerDied","Data":"2d0182fb84b8841913b5e519c4275a805d1ab67d27a93b23760546b6f43b887b"} Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.489204 4829 scope.go:117] "RemoveContainer" containerID="4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.515344 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qv75"] Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.519406 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7qv75"] Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.526179 4829 scope.go:117] "RemoveContainer" containerID="613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.568882 4829 scope.go:117] "RemoveContainer" containerID="28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105" Oct 02 07:30:15 crc kubenswrapper[4829]: E1002 07:30:15.569541 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105\": container with ID starting with 28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105 not found: ID does not exist" containerID="28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.569603 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105"} err="failed to get container status \"28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105\": rpc error: code = NotFound desc = could not find container \"28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105\": container with ID starting with 28d161f9d9e968fa78c6429241bc6d407ddf6adc95fbb18e72a098a47c377105 not found: ID does not exist" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.569645 4829 scope.go:117] "RemoveContainer" containerID="4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1" Oct 02 07:30:15 crc kubenswrapper[4829]: E1002 07:30:15.570294 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1\": container with ID starting with 4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1 not found: ID does not exist" containerID="4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.570339 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1"} err="failed to get container status \"4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1\": rpc error: code = NotFound desc = could not find container \"4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1\": container with ID starting with 4e15728cb3ed9408110f6cbbddca5ad8303d1cf597af0b99b6063b9a5c06faf1 not found: ID does not exist" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.570414 4829 scope.go:117] "RemoveContainer" containerID="613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613" Oct 02 07:30:15 crc kubenswrapper[4829]: E1002 07:30:15.570903 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613\": container with ID starting with 613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613 not found: ID does not exist" containerID="613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613" Oct 02 07:30:15 crc kubenswrapper[4829]: I1002 07:30:15.570947 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613"} err="failed to get container status \"613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613\": rpc error: code = NotFound desc = could not find container \"613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613\": container with ID starting with 613a4550a1f4c7d96b10987c1c0f6015664640b789e112ae5ffaf6f6400ef613 not found: ID does not exist" Oct 02 07:30:16 crc kubenswrapper[4829]: I1002 07:30:16.474496 4829 generic.go:334] "Generic (PLEG): container finished" podID="a85bf9a0-7c63-492e-8662-c463be70aad0" containerID="1abe2c5c3e2989e829a75a2f45e7bd76f651961bdd81d96369b60e8f07be75d7" exitCode=0 Oct 02 07:30:16 crc kubenswrapper[4829]: I1002 07:30:16.474565 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerDied","Data":"1abe2c5c3e2989e829a75a2f45e7bd76f651961bdd81d96369b60e8f07be75d7"} Oct 02 07:30:17 crc kubenswrapper[4829]: I1002 07:30:17.470933 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" path="/var/lib/kubelet/pods/e7b475c1-50d5-4956-b9f4-dd90f360da8f/volumes" Oct 02 07:30:17 crc kubenswrapper[4829]: I1002 07:30:17.486928 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"9339219cc8d781a8c49017a1495e5cbaf831d1471d6b4664890f4f1b74e6f2c4"} Oct 02 07:30:17 crc kubenswrapper[4829]: I1002 07:30:17.486962 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"35c30fb2e341660797e64818ddc259b61f32510d709577c35ef890b787742d43"} Oct 02 07:30:17 crc kubenswrapper[4829]: I1002 07:30:17.486972 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"612e6d627077c23fdceede6e48ea650c54be20ebf4ebad64935154c539ca93de"} Oct 02 07:30:17 crc kubenswrapper[4829]: I1002 07:30:17.486981 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"f49843969d856a71ddbd60b4136030072320d6a71f3cfd3cb6c8f132fceaa033"} Oct 02 07:30:17 crc kubenswrapper[4829]: I1002 07:30:17.486990 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"ae4f029be7b55b009e4a9002bb0e7d343a93d1409843d23cf44a83bfbb9644c0"} Oct 02 07:30:18 crc kubenswrapper[4829]: I1002 07:30:18.319036 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-mvlxd" Oct 02 07:30:18 crc kubenswrapper[4829]: I1002 07:30:18.505606 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8sxd9" event={"ID":"a85bf9a0-7c63-492e-8662-c463be70aad0","Type":"ContainerStarted","Data":"51283fabe00acc4475c25ea51f4ec0e01cb9a59732b4b8fa61f8ac79a21af2b6"} Oct 02 07:30:18 crc kubenswrapper[4829]: I1002 07:30:18.505865 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:18 crc kubenswrapper[4829]: I1002 07:30:18.535191 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-8sxd9" podStartSLOduration=5.947758465 podStartE2EDuration="14.535165068s" podCreationTimestamp="2025-10-02 07:30:04 +0000 UTC" firstStartedPulling="2025-10-02 07:30:04.818094376 +0000 UTC m=+796.157742771" lastFinishedPulling="2025-10-02 07:30:13.405500939 +0000 UTC m=+804.745149374" observedRunningTime="2025-10-02 07:30:18.530786708 +0000 UTC m=+809.870435183" watchObservedRunningTime="2025-10-02 07:30:18.535165068 +0000 UTC m=+809.874813513" Oct 02 07:30:19 crc kubenswrapper[4829]: I1002 07:30:19.668293 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:19 crc kubenswrapper[4829]: I1002 07:30:19.706748 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.157919 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n95gp"] Oct 02 07:30:21 crc kubenswrapper[4829]: E1002 07:30:21.158449 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="extract-utilities" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158466 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="extract-utilities" Oct 02 07:30:21 crc kubenswrapper[4829]: E1002 07:30:21.158481 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="extract-utilities" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158488 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="extract-utilities" Oct 02 07:30:21 crc kubenswrapper[4829]: E1002 07:30:21.158500 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="extract-content" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158507 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="extract-content" Oct 02 07:30:21 crc kubenswrapper[4829]: E1002 07:30:21.158527 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="registry-server" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158535 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="registry-server" Oct 02 07:30:21 crc kubenswrapper[4829]: E1002 07:30:21.158550 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="registry-server" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158557 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="registry-server" Oct 02 07:30:21 crc kubenswrapper[4829]: E1002 07:30:21.158572 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="extract-content" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158579 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="extract-content" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158711 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8d76d5c-3014-461e-8bac-253a85703dd9" containerName="registry-server" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.158728 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b475c1-50d5-4956-b9f4-dd90f360da8f" containerName="registry-server" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.162984 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.171578 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n95gp"] Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.350042 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhflz\" (UniqueName: \"kubernetes.io/projected/d0590b92-f618-4e54-aea6-66c88f9f0f4a-kube-api-access-hhflz\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.350098 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0590b92-f618-4e54-aea6-66c88f9f0f4a-catalog-content\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.350140 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0590b92-f618-4e54-aea6-66c88f9f0f4a-utilities\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.451288 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0590b92-f618-4e54-aea6-66c88f9f0f4a-catalog-content\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.451367 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0590b92-f618-4e54-aea6-66c88f9f0f4a-utilities\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.451424 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhflz\" (UniqueName: \"kubernetes.io/projected/d0590b92-f618-4e54-aea6-66c88f9f0f4a-kube-api-access-hhflz\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.452032 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0590b92-f618-4e54-aea6-66c88f9f0f4a-catalog-content\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.452292 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0590b92-f618-4e54-aea6-66c88f9f0f4a-utilities\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.478246 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhflz\" (UniqueName: \"kubernetes.io/projected/d0590b92-f618-4e54-aea6-66c88f9f0f4a-kube-api-access-hhflz\") pod \"community-operators-n95gp\" (UID: \"d0590b92-f618-4e54-aea6-66c88f9f0f4a\") " pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.479520 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:21 crc kubenswrapper[4829]: I1002 07:30:21.922286 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n95gp"] Oct 02 07:30:21 crc kubenswrapper[4829]: W1002 07:30:21.926031 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0590b92_f618_4e54_aea6_66c88f9f0f4a.slice/crio-9a8d35e7a24eaacfb97753fce796c80047e854e6114759ebe213549ea598fd0d WatchSource:0}: Error finding container 9a8d35e7a24eaacfb97753fce796c80047e854e6114759ebe213549ea598fd0d: Status 404 returned error can't find the container with id 9a8d35e7a24eaacfb97753fce796c80047e854e6114759ebe213549ea598fd0d Oct 02 07:30:22 crc kubenswrapper[4829]: I1002 07:30:22.541093 4829 generic.go:334] "Generic (PLEG): container finished" podID="d0590b92-f618-4e54-aea6-66c88f9f0f4a" containerID="8d54bc3e3e80707d4fc239bd0a94a5a37a07be14539c0c25595ffabc1bcd9a0e" exitCode=0 Oct 02 07:30:22 crc kubenswrapper[4829]: I1002 07:30:22.541315 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n95gp" event={"ID":"d0590b92-f618-4e54-aea6-66c88f9f0f4a","Type":"ContainerDied","Data":"8d54bc3e3e80707d4fc239bd0a94a5a37a07be14539c0c25595ffabc1bcd9a0e"} Oct 02 07:30:22 crc kubenswrapper[4829]: I1002 07:30:22.541548 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n95gp" event={"ID":"d0590b92-f618-4e54-aea6-66c88f9f0f4a","Type":"ContainerStarted","Data":"9a8d35e7a24eaacfb97753fce796c80047e854e6114759ebe213549ea598fd0d"} Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.538175 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2kptc"] Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.540452 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.545815 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2kptc"] Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.549687 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-6g4mv" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.549916 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.550078 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.606308 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhzzb\" (UniqueName: \"kubernetes.io/projected/ec78701c-a17d-4858-b45d-395e8ba9fdb5-kube-api-access-xhzzb\") pod \"openstack-operator-index-2kptc\" (UID: \"ec78701c-a17d-4858-b45d-395e8ba9fdb5\") " pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.704422 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-t78d2" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.707122 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhzzb\" (UniqueName: \"kubernetes.io/projected/ec78701c-a17d-4858-b45d-395e8ba9fdb5-kube-api-access-xhzzb\") pod \"openstack-operator-index-2kptc\" (UID: \"ec78701c-a17d-4858-b45d-395e8ba9fdb5\") " pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.744517 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-mpn22" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.750129 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhzzb\" (UniqueName: \"kubernetes.io/projected/ec78701c-a17d-4858-b45d-395e8ba9fdb5-kube-api-access-xhzzb\") pod \"openstack-operator-index-2kptc\" (UID: \"ec78701c-a17d-4858-b45d-395e8ba9fdb5\") " pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:24 crc kubenswrapper[4829]: I1002 07:30:24.875681 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:26 crc kubenswrapper[4829]: I1002 07:30:26.568907 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n95gp" event={"ID":"d0590b92-f618-4e54-aea6-66c88f9f0f4a","Type":"ContainerStarted","Data":"8b9d002a15e226e4ca4fd901cd41e1167ba42b2df43f3e11dcd6ce2ca4612b0e"} Oct 02 07:30:26 crc kubenswrapper[4829]: I1002 07:30:26.653142 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2kptc"] Oct 02 07:30:26 crc kubenswrapper[4829]: W1002 07:30:26.724323 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec78701c_a17d_4858_b45d_395e8ba9fdb5.slice/crio-d05d7a71c24823b93e800c47b632d705c70aa35e345926daa339384e2c7437ba WatchSource:0}: Error finding container d05d7a71c24823b93e800c47b632d705c70aa35e345926daa339384e2c7437ba: Status 404 returned error can't find the container with id d05d7a71c24823b93e800c47b632d705c70aa35e345926daa339384e2c7437ba Oct 02 07:30:27 crc kubenswrapper[4829]: I1002 07:30:27.577801 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2kptc" event={"ID":"ec78701c-a17d-4858-b45d-395e8ba9fdb5","Type":"ContainerStarted","Data":"d05d7a71c24823b93e800c47b632d705c70aa35e345926daa339384e2c7437ba"} Oct 02 07:30:27 crc kubenswrapper[4829]: I1002 07:30:27.580813 4829 generic.go:334] "Generic (PLEG): container finished" podID="d0590b92-f618-4e54-aea6-66c88f9f0f4a" containerID="8b9d002a15e226e4ca4fd901cd41e1167ba42b2df43f3e11dcd6ce2ca4612b0e" exitCode=0 Oct 02 07:30:27 crc kubenswrapper[4829]: I1002 07:30:27.580853 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n95gp" event={"ID":"d0590b92-f618-4e54-aea6-66c88f9f0f4a","Type":"ContainerDied","Data":"8b9d002a15e226e4ca4fd901cd41e1167ba42b2df43f3e11dcd6ce2ca4612b0e"} Oct 02 07:30:29 crc kubenswrapper[4829]: I1002 07:30:29.601435 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n95gp" event={"ID":"d0590b92-f618-4e54-aea6-66c88f9f0f4a","Type":"ContainerStarted","Data":"5f5f7bacdcabfb7ab4b35583365e5647805cd3497d802b02b57d62d686d1639e"} Oct 02 07:30:29 crc kubenswrapper[4829]: I1002 07:30:29.605623 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2kptc" event={"ID":"ec78701c-a17d-4858-b45d-395e8ba9fdb5","Type":"ContainerStarted","Data":"27fb4238138afbee6897b7334c43fc04df8e9026403e74b6140b62b14c33e84d"} Oct 02 07:30:29 crc kubenswrapper[4829]: I1002 07:30:29.623090 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n95gp" podStartSLOduration=1.848051387 podStartE2EDuration="8.623065245s" podCreationTimestamp="2025-10-02 07:30:21 +0000 UTC" firstStartedPulling="2025-10-02 07:30:22.543580915 +0000 UTC m=+813.883229360" lastFinishedPulling="2025-10-02 07:30:29.318594763 +0000 UTC m=+820.658243218" observedRunningTime="2025-10-02 07:30:29.620702169 +0000 UTC m=+820.960350624" watchObservedRunningTime="2025-10-02 07:30:29.623065245 +0000 UTC m=+820.962713690" Oct 02 07:30:29 crc kubenswrapper[4829]: I1002 07:30:29.647698 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2kptc" podStartSLOduration=3.042697628 podStartE2EDuration="5.647667826s" podCreationTimestamp="2025-10-02 07:30:24 +0000 UTC" firstStartedPulling="2025-10-02 07:30:26.734861477 +0000 UTC m=+818.074509882" lastFinishedPulling="2025-10-02 07:30:29.339831635 +0000 UTC m=+820.679480080" observedRunningTime="2025-10-02 07:30:29.643887554 +0000 UTC m=+820.983535959" watchObservedRunningTime="2025-10-02 07:30:29.647667826 +0000 UTC m=+820.987316251" Oct 02 07:30:29 crc kubenswrapper[4829]: I1002 07:30:29.935666 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2kptc"] Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.542656 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-bqsd6"] Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.544353 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.556057 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bqsd6"] Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.622186 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhbf5\" (UniqueName: \"kubernetes.io/projected/fba8d39c-ec14-48e6-a9de-41f33e873f09-kube-api-access-nhbf5\") pod \"openstack-operator-index-bqsd6\" (UID: \"fba8d39c-ec14-48e6-a9de-41f33e873f09\") " pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.724524 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhbf5\" (UniqueName: \"kubernetes.io/projected/fba8d39c-ec14-48e6-a9de-41f33e873f09-kube-api-access-nhbf5\") pod \"openstack-operator-index-bqsd6\" (UID: \"fba8d39c-ec14-48e6-a9de-41f33e873f09\") " pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.762961 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhbf5\" (UniqueName: \"kubernetes.io/projected/fba8d39c-ec14-48e6-a9de-41f33e873f09-kube-api-access-nhbf5\") pod \"openstack-operator-index-bqsd6\" (UID: \"fba8d39c-ec14-48e6-a9de-41f33e873f09\") " pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:30 crc kubenswrapper[4829]: I1002 07:30:30.875270 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:31 crc kubenswrapper[4829]: I1002 07:30:31.120056 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bqsd6"] Oct 02 07:30:31 crc kubenswrapper[4829]: I1002 07:30:31.480380 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:31 crc kubenswrapper[4829]: I1002 07:30:31.480988 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:31 crc kubenswrapper[4829]: I1002 07:30:31.620890 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bqsd6" event={"ID":"fba8d39c-ec14-48e6-a9de-41f33e873f09","Type":"ContainerStarted","Data":"5eda1bf28af8f035310bd0f8c58d302ef25f77a12b67ae07d6c88a76088df78e"} Oct 02 07:30:31 crc kubenswrapper[4829]: I1002 07:30:31.621012 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2kptc" podUID="ec78701c-a17d-4858-b45d-395e8ba9fdb5" containerName="registry-server" containerID="cri-o://27fb4238138afbee6897b7334c43fc04df8e9026403e74b6140b62b14c33e84d" gracePeriod=2 Oct 02 07:30:32 crc kubenswrapper[4829]: I1002 07:30:32.557277 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n95gp" podUID="d0590b92-f618-4e54-aea6-66c88f9f0f4a" containerName="registry-server" probeResult="failure" output=< Oct 02 07:30:32 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:30:32 crc kubenswrapper[4829]: > Oct 02 07:30:32 crc kubenswrapper[4829]: I1002 07:30:32.628367 4829 generic.go:334] "Generic (PLEG): container finished" podID="ec78701c-a17d-4858-b45d-395e8ba9fdb5" containerID="27fb4238138afbee6897b7334c43fc04df8e9026403e74b6140b62b14c33e84d" exitCode=0 Oct 02 07:30:32 crc kubenswrapper[4829]: I1002 07:30:32.628457 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2kptc" event={"ID":"ec78701c-a17d-4858-b45d-395e8ba9fdb5","Type":"ContainerDied","Data":"27fb4238138afbee6897b7334c43fc04df8e9026403e74b6140b62b14c33e84d"} Oct 02 07:30:33 crc kubenswrapper[4829]: I1002 07:30:33.942202 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.076800 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhzzb\" (UniqueName: \"kubernetes.io/projected/ec78701c-a17d-4858-b45d-395e8ba9fdb5-kube-api-access-xhzzb\") pod \"ec78701c-a17d-4858-b45d-395e8ba9fdb5\" (UID: \"ec78701c-a17d-4858-b45d-395e8ba9fdb5\") " Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.085278 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec78701c-a17d-4858-b45d-395e8ba9fdb5-kube-api-access-xhzzb" (OuterVolumeSpecName: "kube-api-access-xhzzb") pod "ec78701c-a17d-4858-b45d-395e8ba9fdb5" (UID: "ec78701c-a17d-4858-b45d-395e8ba9fdb5"). InnerVolumeSpecName "kube-api-access-xhzzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.179206 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhzzb\" (UniqueName: \"kubernetes.io/projected/ec78701c-a17d-4858-b45d-395e8ba9fdb5-kube-api-access-xhzzb\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.646890 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bqsd6" event={"ID":"fba8d39c-ec14-48e6-a9de-41f33e873f09","Type":"ContainerStarted","Data":"4020c6eb161874c8d14b3229d8cfa6e9211e86faa202692d055599543d0c9f34"} Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.648858 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2kptc" event={"ID":"ec78701c-a17d-4858-b45d-395e8ba9fdb5","Type":"ContainerDied","Data":"d05d7a71c24823b93e800c47b632d705c70aa35e345926daa339384e2c7437ba"} Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.648902 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2kptc" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.648940 4829 scope.go:117] "RemoveContainer" containerID="27fb4238138afbee6897b7334c43fc04df8e9026403e74b6140b62b14c33e84d" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.674893 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-bqsd6" podStartSLOduration=1.7036520880000001 podStartE2EDuration="4.674870223s" podCreationTimestamp="2025-10-02 07:30:30 +0000 UTC" firstStartedPulling="2025-10-02 07:30:31.129273894 +0000 UTC m=+822.468922289" lastFinishedPulling="2025-10-02 07:30:34.100491979 +0000 UTC m=+825.440140424" observedRunningTime="2025-10-02 07:30:34.671642738 +0000 UTC m=+826.011291173" watchObservedRunningTime="2025-10-02 07:30:34.674870223 +0000 UTC m=+826.014518658" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.682213 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-8sxd9" Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.745870 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-2kptc"] Oct 02 07:30:34 crc kubenswrapper[4829]: I1002 07:30:34.749979 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-2kptc"] Oct 02 07:30:35 crc kubenswrapper[4829]: I1002 07:30:35.472094 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec78701c-a17d-4858-b45d-395e8ba9fdb5" path="/var/lib/kubelet/pods/ec78701c-a17d-4858-b45d-395e8ba9fdb5/volumes" Oct 02 07:30:40 crc kubenswrapper[4829]: I1002 07:30:40.875221 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:40 crc kubenswrapper[4829]: I1002 07:30:40.875755 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:40 crc kubenswrapper[4829]: I1002 07:30:40.933893 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:41 crc kubenswrapper[4829]: I1002 07:30:41.554113 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:41 crc kubenswrapper[4829]: I1002 07:30:41.635982 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n95gp" Oct 02 07:30:41 crc kubenswrapper[4829]: I1002 07:30:41.724816 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n95gp"] Oct 02 07:30:41 crc kubenswrapper[4829]: I1002 07:30:41.758405 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-bqsd6" Oct 02 07:30:41 crc kubenswrapper[4829]: I1002 07:30:41.811526 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9jpj"] Oct 02 07:30:41 crc kubenswrapper[4829]: I1002 07:30:41.811876 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j9jpj" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="registry-server" containerID="cri-o://8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad" gracePeriod=2 Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.277154 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.355369 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-utilities\") pod \"b3d721af-feab-49f1-833f-fcf222170ce3\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.355466 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptdw4\" (UniqueName: \"kubernetes.io/projected/b3d721af-feab-49f1-833f-fcf222170ce3-kube-api-access-ptdw4\") pod \"b3d721af-feab-49f1-833f-fcf222170ce3\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.355566 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-catalog-content\") pod \"b3d721af-feab-49f1-833f-fcf222170ce3\" (UID: \"b3d721af-feab-49f1-833f-fcf222170ce3\") " Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.358077 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-utilities" (OuterVolumeSpecName: "utilities") pod "b3d721af-feab-49f1-833f-fcf222170ce3" (UID: "b3d721af-feab-49f1-833f-fcf222170ce3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.366430 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d721af-feab-49f1-833f-fcf222170ce3-kube-api-access-ptdw4" (OuterVolumeSpecName: "kube-api-access-ptdw4") pod "b3d721af-feab-49f1-833f-fcf222170ce3" (UID: "b3d721af-feab-49f1-833f-fcf222170ce3"). InnerVolumeSpecName "kube-api-access-ptdw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.409479 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3d721af-feab-49f1-833f-fcf222170ce3" (UID: "b3d721af-feab-49f1-833f-fcf222170ce3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.456741 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.456804 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d721af-feab-49f1-833f-fcf222170ce3-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.456817 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptdw4\" (UniqueName: \"kubernetes.io/projected/b3d721af-feab-49f1-833f-fcf222170ce3-kube-api-access-ptdw4\") on node \"crc\" DevicePath \"\"" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.726096 4829 generic.go:334] "Generic (PLEG): container finished" podID="b3d721af-feab-49f1-833f-fcf222170ce3" containerID="8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad" exitCode=0 Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.726175 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9jpj" event={"ID":"b3d721af-feab-49f1-833f-fcf222170ce3","Type":"ContainerDied","Data":"8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad"} Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.726212 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j9jpj" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.726287 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j9jpj" event={"ID":"b3d721af-feab-49f1-833f-fcf222170ce3","Type":"ContainerDied","Data":"a2a736fce4d1b8cbf9a532317d10dc95f60ce5231f7ac953b3ea15a5db3a2449"} Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.726330 4829 scope.go:117] "RemoveContainer" containerID="8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.751626 4829 scope.go:117] "RemoveContainer" containerID="43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.781437 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j9jpj"] Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.786412 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j9jpj"] Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.787451 4829 scope.go:117] "RemoveContainer" containerID="ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.820655 4829 scope.go:117] "RemoveContainer" containerID="8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad" Oct 02 07:30:42 crc kubenswrapper[4829]: E1002 07:30:42.821203 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad\": container with ID starting with 8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad not found: ID does not exist" containerID="8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.821270 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad"} err="failed to get container status \"8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad\": rpc error: code = NotFound desc = could not find container \"8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad\": container with ID starting with 8784de52ecd837ed360410da44616a5ce985c2d7150dde362e2950c501159dad not found: ID does not exist" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.821305 4829 scope.go:117] "RemoveContainer" containerID="43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3" Oct 02 07:30:42 crc kubenswrapper[4829]: E1002 07:30:42.821796 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3\": container with ID starting with 43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3 not found: ID does not exist" containerID="43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.821844 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3"} err="failed to get container status \"43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3\": rpc error: code = NotFound desc = could not find container \"43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3\": container with ID starting with 43ea0a13d2b2b094f6b5b10d19745d94467d72f2aa2be67070099bd621ad60c3 not found: ID does not exist" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.821875 4829 scope.go:117] "RemoveContainer" containerID="ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2" Oct 02 07:30:42 crc kubenswrapper[4829]: E1002 07:30:42.822297 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2\": container with ID starting with ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2 not found: ID does not exist" containerID="ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2" Oct 02 07:30:42 crc kubenswrapper[4829]: I1002 07:30:42.822339 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2"} err="failed to get container status \"ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2\": rpc error: code = NotFound desc = could not find container \"ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2\": container with ID starting with ec0ec816bfc7b0e163b10af44f4667bb75f94206a38da23427dcfa7f690915b2 not found: ID does not exist" Oct 02 07:30:43 crc kubenswrapper[4829]: I1002 07:30:43.470314 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" path="/var/lib/kubelet/pods/b3d721af-feab-49f1-833f-fcf222170ce3/volumes" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.791183 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5"] Oct 02 07:30:54 crc kubenswrapper[4829]: E1002 07:30:54.792247 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="registry-server" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.792269 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="registry-server" Oct 02 07:30:54 crc kubenswrapper[4829]: E1002 07:30:54.792290 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec78701c-a17d-4858-b45d-395e8ba9fdb5" containerName="registry-server" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.792302 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec78701c-a17d-4858-b45d-395e8ba9fdb5" containerName="registry-server" Oct 02 07:30:54 crc kubenswrapper[4829]: E1002 07:30:54.792317 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="extract-utilities" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.792331 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="extract-utilities" Oct 02 07:30:54 crc kubenswrapper[4829]: E1002 07:30:54.792349 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="extract-content" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.792361 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="extract-content" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.792576 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d721af-feab-49f1-833f-fcf222170ce3" containerName="registry-server" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.792602 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec78701c-a17d-4858-b45d-395e8ba9fdb5" containerName="registry-server" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.794021 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.796269 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zsxqx" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.798323 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5"] Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.840908 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-util\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.841053 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-bundle\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.841146 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvtc8\" (UniqueName: \"kubernetes.io/projected/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-kube-api-access-nvtc8\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.942397 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvtc8\" (UniqueName: \"kubernetes.io/projected/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-kube-api-access-nvtc8\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.942586 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-util\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.942862 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-bundle\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.943414 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-util\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.943492 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-bundle\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:54 crc kubenswrapper[4829]: I1002 07:30:54.971821 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvtc8\" (UniqueName: \"kubernetes.io/projected/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-kube-api-access-nvtc8\") pod \"92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:55 crc kubenswrapper[4829]: I1002 07:30:55.117625 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:30:55 crc kubenswrapper[4829]: I1002 07:30:55.635456 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5"] Oct 02 07:30:55 crc kubenswrapper[4829]: W1002 07:30:55.643512 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c06d3e9_2ef1_4e91_b24d_89cfeecf485e.slice/crio-fce720e7fdd8c0c4b3069d2ca42ee6d5af24454262d37cde55e728b56b3ad703 WatchSource:0}: Error finding container fce720e7fdd8c0c4b3069d2ca42ee6d5af24454262d37cde55e728b56b3ad703: Status 404 returned error can't find the container with id fce720e7fdd8c0c4b3069d2ca42ee6d5af24454262d37cde55e728b56b3ad703 Oct 02 07:30:55 crc kubenswrapper[4829]: I1002 07:30:55.829779 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" event={"ID":"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e","Type":"ContainerStarted","Data":"ab38051ec2564502fb16547b3e194cb6777ba4212514fa76aa51db99b1f31bae"} Oct 02 07:30:55 crc kubenswrapper[4829]: I1002 07:30:55.829846 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" event={"ID":"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e","Type":"ContainerStarted","Data":"fce720e7fdd8c0c4b3069d2ca42ee6d5af24454262d37cde55e728b56b3ad703"} Oct 02 07:30:56 crc kubenswrapper[4829]: I1002 07:30:56.839985 4829 generic.go:334] "Generic (PLEG): container finished" podID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerID="ab38051ec2564502fb16547b3e194cb6777ba4212514fa76aa51db99b1f31bae" exitCode=0 Oct 02 07:30:56 crc kubenswrapper[4829]: I1002 07:30:56.840042 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" event={"ID":"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e","Type":"ContainerDied","Data":"ab38051ec2564502fb16547b3e194cb6777ba4212514fa76aa51db99b1f31bae"} Oct 02 07:30:57 crc kubenswrapper[4829]: I1002 07:30:57.851077 4829 generic.go:334] "Generic (PLEG): container finished" podID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerID="b9046225de933dcbfe7c01f9c826e9e5e5174bcae3a1266eb44a7ff5bce59c8f" exitCode=0 Oct 02 07:30:57 crc kubenswrapper[4829]: I1002 07:30:57.851163 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" event={"ID":"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e","Type":"ContainerDied","Data":"b9046225de933dcbfe7c01f9c826e9e5e5174bcae3a1266eb44a7ff5bce59c8f"} Oct 02 07:30:58 crc kubenswrapper[4829]: I1002 07:30:58.865845 4829 generic.go:334] "Generic (PLEG): container finished" podID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerID="0fc57ca48ff3b8498e8b7f4b9462728939853b2024b9c4d7735a2e72f5e2b539" exitCode=0 Oct 02 07:30:58 crc kubenswrapper[4829]: I1002 07:30:58.865906 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" event={"ID":"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e","Type":"ContainerDied","Data":"0fc57ca48ff3b8498e8b7f4b9462728939853b2024b9c4d7735a2e72f5e2b539"} Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.194480 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.371971 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-util\") pod \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.372331 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-bundle\") pod \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.372422 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvtc8\" (UniqueName: \"kubernetes.io/projected/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-kube-api-access-nvtc8\") pod \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\" (UID: \"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e\") " Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.373386 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-bundle" (OuterVolumeSpecName: "bundle") pod "1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" (UID: "1c06d3e9-2ef1-4e91-b24d-89cfeecf485e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.373660 4829 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.382611 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-kube-api-access-nvtc8" (OuterVolumeSpecName: "kube-api-access-nvtc8") pod "1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" (UID: "1c06d3e9-2ef1-4e91-b24d-89cfeecf485e"). InnerVolumeSpecName "kube-api-access-nvtc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.412411 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-util" (OuterVolumeSpecName: "util") pod "1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" (UID: "1c06d3e9-2ef1-4e91-b24d-89cfeecf485e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.475368 4829 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-util\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.475400 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvtc8\" (UniqueName: \"kubernetes.io/projected/1c06d3e9-2ef1-4e91-b24d-89cfeecf485e-kube-api-access-nvtc8\") on node \"crc\" DevicePath \"\"" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.888070 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" event={"ID":"1c06d3e9-2ef1-4e91-b24d-89cfeecf485e","Type":"ContainerDied","Data":"fce720e7fdd8c0c4b3069d2ca42ee6d5af24454262d37cde55e728b56b3ad703"} Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.888140 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fce720e7fdd8c0c4b3069d2ca42ee6d5af24454262d37cde55e728b56b3ad703" Oct 02 07:31:00 crc kubenswrapper[4829]: I1002 07:31:00.888219 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.128099 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h"] Oct 02 07:31:03 crc kubenswrapper[4829]: E1002 07:31:03.128738 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="extract" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.128759 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="extract" Oct 02 07:31:03 crc kubenswrapper[4829]: E1002 07:31:03.128783 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="pull" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.128795 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="pull" Oct 02 07:31:03 crc kubenswrapper[4829]: E1002 07:31:03.128833 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="util" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.128846 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="util" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.129036 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c06d3e9-2ef1-4e91-b24d-89cfeecf485e" containerName="extract" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.130104 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.133147 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-pbmgs" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.166966 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h"] Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.315387 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w8ns\" (UniqueName: \"kubernetes.io/projected/507314d1-5e67-4900-99bd-fd445866f35b-kube-api-access-2w8ns\") pod \"openstack-operator-controller-operator-6f46d87668-x4x8h\" (UID: \"507314d1-5e67-4900-99bd-fd445866f35b\") " pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.416631 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w8ns\" (UniqueName: \"kubernetes.io/projected/507314d1-5e67-4900-99bd-fd445866f35b-kube-api-access-2w8ns\") pod \"openstack-operator-controller-operator-6f46d87668-x4x8h\" (UID: \"507314d1-5e67-4900-99bd-fd445866f35b\") " pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.435430 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w8ns\" (UniqueName: \"kubernetes.io/projected/507314d1-5e67-4900-99bd-fd445866f35b-kube-api-access-2w8ns\") pod \"openstack-operator-controller-operator-6f46d87668-x4x8h\" (UID: \"507314d1-5e67-4900-99bd-fd445866f35b\") " pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.464766 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:03 crc kubenswrapper[4829]: I1002 07:31:03.952740 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h"] Oct 02 07:31:04 crc kubenswrapper[4829]: I1002 07:31:04.913716 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" event={"ID":"507314d1-5e67-4900-99bd-fd445866f35b","Type":"ContainerStarted","Data":"8118eb346cbfe71427c987f4754a45f273dd53d34fa3d1cd76fefc4bc7f49c03"} Oct 02 07:31:07 crc kubenswrapper[4829]: I1002 07:31:07.930579 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" event={"ID":"507314d1-5e67-4900-99bd-fd445866f35b","Type":"ContainerStarted","Data":"ff52814f7e0d129d80ca42386534b7b1d2e6fd3dadff3a80c4ea7084d56e9432"} Oct 02 07:31:09 crc kubenswrapper[4829]: I1002 07:31:09.945039 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" event={"ID":"507314d1-5e67-4900-99bd-fd445866f35b","Type":"ContainerStarted","Data":"f48597a70e17c566f10c9aa18b38b2dda4fc464b2903ea9bec3ddb858bcd217e"} Oct 02 07:31:09 crc kubenswrapper[4829]: I1002 07:31:09.945456 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:09 crc kubenswrapper[4829]: I1002 07:31:09.978497 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" podStartSLOduration=1.370938486 podStartE2EDuration="6.978477387s" podCreationTimestamp="2025-10-02 07:31:03 +0000 UTC" firstStartedPulling="2025-10-02 07:31:03.963476286 +0000 UTC m=+855.303124691" lastFinishedPulling="2025-10-02 07:31:09.571015177 +0000 UTC m=+860.910663592" observedRunningTime="2025-10-02 07:31:09.974396316 +0000 UTC m=+861.314044731" watchObservedRunningTime="2025-10-02 07:31:09.978477387 +0000 UTC m=+861.318125792" Oct 02 07:31:13 crc kubenswrapper[4829]: I1002 07:31:13.473952 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6f46d87668-x4x8h" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.311777 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.313057 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.319255 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.320155 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-xs6hw" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.320987 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.326555 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-dzvk5" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.330022 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.330988 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.332871 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-dkx8c" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.349120 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.358607 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.359811 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.363937 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5bfst" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.367197 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.374110 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.389015 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt4l6\" (UniqueName: \"kubernetes.io/projected/f588d058-d57b-4824-b59d-40731650907f-kube-api-access-xt4l6\") pod \"cinder-operator-controller-manager-644bddb6d8-7wrzt\" (UID: \"f588d058-d57b-4824-b59d-40731650907f\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.389064 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx7t9\" (UniqueName: \"kubernetes.io/projected/349c96b5-e59f-4eac-8646-e5459775391c-kube-api-access-qx7t9\") pod \"designate-operator-controller-manager-84f4f7b77b-rs8gs\" (UID: \"349c96b5-e59f-4eac-8646-e5459775391c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.389117 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpj8q\" (UniqueName: \"kubernetes.io/projected/104543c5-16f7-47d1-b2cb-dedd729a87b7-kube-api-access-mpj8q\") pod \"glance-operator-controller-manager-84958c4d49-5r5zb\" (UID: \"104543c5-16f7-47d1-b2cb-dedd729a87b7\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.389139 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z8z8\" (UniqueName: \"kubernetes.io/projected/895426fb-40c5-426b-ab48-f5ac31d22e06-kube-api-access-4z8z8\") pod \"barbican-operator-controller-manager-6ff8b75857-nf7gj\" (UID: \"895426fb-40c5-426b-ab48-f5ac31d22e06\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.396195 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.403275 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.404457 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.409513 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-f8d8x" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.411415 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.412489 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.416008 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.416195 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-k59ks" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.416306 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.417262 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.419876 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.420167 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-97x4b" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.444927 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.446023 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.449037 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-mgs2r" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.456153 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.457459 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.461658 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6sl5p" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.480434 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.480476 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495019 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt4l6\" (UniqueName: \"kubernetes.io/projected/f588d058-d57b-4824-b59d-40731650907f-kube-api-access-xt4l6\") pod \"cinder-operator-controller-manager-644bddb6d8-7wrzt\" (UID: \"f588d058-d57b-4824-b59d-40731650907f\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495300 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlppx\" (UniqueName: \"kubernetes.io/projected/e46a78e2-3b2a-47c5-8a9a-03877339340d-kube-api-access-dlppx\") pod \"infra-operator-controller-manager-9d6c5db85-zc9kp\" (UID: \"e46a78e2-3b2a-47c5-8a9a-03877339340d\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495332 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx7t9\" (UniqueName: \"kubernetes.io/projected/349c96b5-e59f-4eac-8646-e5459775391c-kube-api-access-qx7t9\") pod \"designate-operator-controller-manager-84f4f7b77b-rs8gs\" (UID: \"349c96b5-e59f-4eac-8646-e5459775391c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495498 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e46a78e2-3b2a-47c5-8a9a-03877339340d-cert\") pod \"infra-operator-controller-manager-9d6c5db85-zc9kp\" (UID: \"e46a78e2-3b2a-47c5-8a9a-03877339340d\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495545 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d67j\" (UniqueName: \"kubernetes.io/projected/88cfec3f-d6c0-4007-a845-a3c4e56bcb63-kube-api-access-6d67j\") pod \"keystone-operator-controller-manager-5bd55b4bff-pf5r7\" (UID: \"88cfec3f-d6c0-4007-a845-a3c4e56bcb63\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495621 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gslcx\" (UniqueName: \"kubernetes.io/projected/5607af74-7a80-4b1c-9b33-690d3f7fe974-kube-api-access-gslcx\") pod \"horizon-operator-controller-manager-9f4696d94-cw4k8\" (UID: \"5607af74-7a80-4b1c-9b33-690d3f7fe974\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495660 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjwzc\" (UniqueName: \"kubernetes.io/projected/6f685eab-a458-4058-9453-5c4a389758e0-kube-api-access-bjwzc\") pod \"heat-operator-controller-manager-5d889d78cf-ztngc\" (UID: \"6f685eab-a458-4058-9453-5c4a389758e0\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495688 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpj8q\" (UniqueName: \"kubernetes.io/projected/104543c5-16f7-47d1-b2cb-dedd729a87b7-kube-api-access-mpj8q\") pod \"glance-operator-controller-manager-84958c4d49-5r5zb\" (UID: \"104543c5-16f7-47d1-b2cb-dedd729a87b7\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495720 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z8z8\" (UniqueName: \"kubernetes.io/projected/895426fb-40c5-426b-ab48-f5ac31d22e06-kube-api-access-4z8z8\") pod \"barbican-operator-controller-manager-6ff8b75857-nf7gj\" (UID: \"895426fb-40c5-426b-ab48-f5ac31d22e06\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.495775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfqch\" (UniqueName: \"kubernetes.io/projected/3f4b21a4-b546-4f69-a432-05ba29b24085-kube-api-access-mfqch\") pod \"ironic-operator-controller-manager-5cd4858477-2njwr\" (UID: \"3f4b21a4-b546-4f69-a432-05ba29b24085\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.505329 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.537108 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx7t9\" (UniqueName: \"kubernetes.io/projected/349c96b5-e59f-4eac-8646-e5459775391c-kube-api-access-qx7t9\") pod \"designate-operator-controller-manager-84f4f7b77b-rs8gs\" (UID: \"349c96b5-e59f-4eac-8646-e5459775391c\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.546795 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.547871 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.549218 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z8z8\" (UniqueName: \"kubernetes.io/projected/895426fb-40c5-426b-ab48-f5ac31d22e06-kube-api-access-4z8z8\") pod \"barbican-operator-controller-manager-6ff8b75857-nf7gj\" (UID: \"895426fb-40c5-426b-ab48-f5ac31d22e06\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.549657 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt4l6\" (UniqueName: \"kubernetes.io/projected/f588d058-d57b-4824-b59d-40731650907f-kube-api-access-xt4l6\") pod \"cinder-operator-controller-manager-644bddb6d8-7wrzt\" (UID: \"f588d058-d57b-4824-b59d-40731650907f\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.549825 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.550131 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-8b22k" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.550966 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpj8q\" (UniqueName: \"kubernetes.io/projected/104543c5-16f7-47d1-b2cb-dedd729a87b7-kube-api-access-mpj8q\") pod \"glance-operator-controller-manager-84958c4d49-5r5zb\" (UID: \"104543c5-16f7-47d1-b2cb-dedd729a87b7\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.577573 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.578537 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.597814 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cprfn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.620167 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlppx\" (UniqueName: \"kubernetes.io/projected/e46a78e2-3b2a-47c5-8a9a-03877339340d-kube-api-access-dlppx\") pod \"infra-operator-controller-manager-9d6c5db85-zc9kp\" (UID: \"e46a78e2-3b2a-47c5-8a9a-03877339340d\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.621493 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e46a78e2-3b2a-47c5-8a9a-03877339340d-cert\") pod \"infra-operator-controller-manager-9d6c5db85-zc9kp\" (UID: \"e46a78e2-3b2a-47c5-8a9a-03877339340d\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.621643 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d67j\" (UniqueName: \"kubernetes.io/projected/88cfec3f-d6c0-4007-a845-a3c4e56bcb63-kube-api-access-6d67j\") pod \"keystone-operator-controller-manager-5bd55b4bff-pf5r7\" (UID: \"88cfec3f-d6c0-4007-a845-a3c4e56bcb63\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.621740 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gslcx\" (UniqueName: \"kubernetes.io/projected/5607af74-7a80-4b1c-9b33-690d3f7fe974-kube-api-access-gslcx\") pod \"horizon-operator-controller-manager-9f4696d94-cw4k8\" (UID: \"5607af74-7a80-4b1c-9b33-690d3f7fe974\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.621813 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjwzc\" (UniqueName: \"kubernetes.io/projected/6f685eab-a458-4058-9453-5c4a389758e0-kube-api-access-bjwzc\") pod \"heat-operator-controller-manager-5d889d78cf-ztngc\" (UID: \"6f685eab-a458-4058-9453-5c4a389758e0\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.621987 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfqch\" (UniqueName: \"kubernetes.io/projected/3f4b21a4-b546-4f69-a432-05ba29b24085-kube-api-access-mfqch\") pod \"ironic-operator-controller-manager-5cd4858477-2njwr\" (UID: \"3f4b21a4-b546-4f69-a432-05ba29b24085\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.622089 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9zg7\" (UniqueName: \"kubernetes.io/projected/3430c81b-20da-4871-9f8c-a408573677ec-kube-api-access-h9zg7\") pod \"mariadb-operator-controller-manager-88c7-p8wc8\" (UID: \"3430c81b-20da-4871-9f8c-a408573677ec\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.622190 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khbjr\" (UniqueName: \"kubernetes.io/projected/c555f81b-a5f9-4b4b-9ae0-5253ebb12099-kube-api-access-khbjr\") pod \"manila-operator-controller-manager-6d68dbc695-tmszn\" (UID: \"c555f81b-a5f9-4b4b-9ae0-5253ebb12099\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.629813 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e46a78e2-3b2a-47c5-8a9a-03877339340d-cert\") pod \"infra-operator-controller-manager-9d6c5db85-zc9kp\" (UID: \"e46a78e2-3b2a-47c5-8a9a-03877339340d\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.641258 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.646936 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.662178 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.682552 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjwzc\" (UniqueName: \"kubernetes.io/projected/6f685eab-a458-4058-9453-5c4a389758e0-kube-api-access-bjwzc\") pod \"heat-operator-controller-manager-5d889d78cf-ztngc\" (UID: \"6f685eab-a458-4058-9453-5c4a389758e0\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.682570 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d67j\" (UniqueName: \"kubernetes.io/projected/88cfec3f-d6c0-4007-a845-a3c4e56bcb63-kube-api-access-6d67j\") pod \"keystone-operator-controller-manager-5bd55b4bff-pf5r7\" (UID: \"88cfec3f-d6c0-4007-a845-a3c4e56bcb63\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.682570 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlppx\" (UniqueName: \"kubernetes.io/projected/e46a78e2-3b2a-47c5-8a9a-03877339340d-kube-api-access-dlppx\") pod \"infra-operator-controller-manager-9d6c5db85-zc9kp\" (UID: \"e46a78e2-3b2a-47c5-8a9a-03877339340d\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.682952 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfqch\" (UniqueName: \"kubernetes.io/projected/3f4b21a4-b546-4f69-a432-05ba29b24085-kube-api-access-mfqch\") pod \"ironic-operator-controller-manager-5cd4858477-2njwr\" (UID: \"3f4b21a4-b546-4f69-a432-05ba29b24085\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.683276 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gslcx\" (UniqueName: \"kubernetes.io/projected/5607af74-7a80-4b1c-9b33-690d3f7fe974-kube-api-access-gslcx\") pod \"horizon-operator-controller-manager-9f4696d94-cw4k8\" (UID: \"5607af74-7a80-4b1c-9b33-690d3f7fe974\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.685764 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.687595 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.688613 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.696748 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-fq9f9" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.699372 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.713536 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.727749 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q6hr\" (UniqueName: \"kubernetes.io/projected/d697f00d-25f4-48fb-808e-1a5fb90ff5fa-kube-api-access-8q6hr\") pod \"neutron-operator-controller-manager-849d5b9b84-2rk2r\" (UID: \"d697f00d-25f4-48fb-808e-1a5fb90ff5fa\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.727832 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9zg7\" (UniqueName: \"kubernetes.io/projected/3430c81b-20da-4871-9f8c-a408573677ec-kube-api-access-h9zg7\") pod \"mariadb-operator-controller-manager-88c7-p8wc8\" (UID: \"3430c81b-20da-4871-9f8c-a408573677ec\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.729013 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.729663 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khbjr\" (UniqueName: \"kubernetes.io/projected/c555f81b-a5f9-4b4b-9ae0-5253ebb12099-kube-api-access-khbjr\") pod \"manila-operator-controller-manager-6d68dbc695-tmszn\" (UID: \"c555f81b-a5f9-4b4b-9ae0-5253ebb12099\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.730372 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.743937 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khbjr\" (UniqueName: \"kubernetes.io/projected/c555f81b-a5f9-4b4b-9ae0-5253ebb12099-kube-api-access-khbjr\") pod \"manila-operator-controller-manager-6d68dbc695-tmszn\" (UID: \"c555f81b-a5f9-4b4b-9ae0-5253ebb12099\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.747206 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.751699 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9zg7\" (UniqueName: \"kubernetes.io/projected/3430c81b-20da-4871-9f8c-a408573677ec-kube-api-access-h9zg7\") pod \"mariadb-operator-controller-manager-88c7-p8wc8\" (UID: \"3430c81b-20da-4871-9f8c-a408573677ec\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.756537 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.758302 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.761731 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-vf8k9" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.763974 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.765074 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.765946 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.768014 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-n9q4x" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.776459 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.791446 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.791845 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.804404 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.805959 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.808153 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.809475 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-bthkh" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.810176 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.811382 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.812950 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-fkxd6" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.819109 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.820069 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.827208 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-5xmnr" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.836035 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q6hr\" (UniqueName: \"kubernetes.io/projected/d697f00d-25f4-48fb-808e-1a5fb90ff5fa-kube-api-access-8q6hr\") pod \"neutron-operator-controller-manager-849d5b9b84-2rk2r\" (UID: \"d697f00d-25f4-48fb-808e-1a5fb90ff5fa\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.837564 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.857853 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q6hr\" (UniqueName: \"kubernetes.io/projected/d697f00d-25f4-48fb-808e-1a5fb90ff5fa-kube-api-access-8q6hr\") pod \"neutron-operator-controller-manager-849d5b9b84-2rk2r\" (UID: \"d697f00d-25f4-48fb-808e-1a5fb90ff5fa\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.863358 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.878770 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.893503 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.909810 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.910869 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.934660 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.934719 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.935883 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.936919 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcwk5\" (UniqueName: \"kubernetes.io/projected/bf88c28f-c14a-4390-83ed-75dc3e41061d-kube-api-access-jcwk5\") pod \"nova-operator-controller-manager-64cd67b5cb-722dn\" (UID: \"bf88c28f-c14a-4390-83ed-75dc3e41061d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.936953 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdhrt\" (UniqueName: \"kubernetes.io/projected/3dee869a-23fe-4926-a561-67a87ca83103-kube-api-access-cdhrt\") pod \"ovn-operator-controller-manager-9976ff44c-5cqsw\" (UID: \"3dee869a-23fe-4926-a561-67a87ca83103\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.937003 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4x6b\" (UniqueName: \"kubernetes.io/projected/2ba71fca-3e78-49e5-bfe5-b831568118ba-kube-api-access-j4x6b\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.937024 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj2hd\" (UniqueName: \"kubernetes.io/projected/c35efaf9-0b1b-4801-81c9-f8db24083049-kube-api-access-zj2hd\") pod \"octavia-operator-controller-manager-7b787867f4-hlq5c\" (UID: \"c35efaf9-0b1b-4801-81c9-f8db24083049\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.937051 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfbw5\" (UniqueName: \"kubernetes.io/projected/8dd30f73-72f3-4733-b5c8-f96753750fb8-kube-api-access-hfbw5\") pod \"placement-operator-controller-manager-589c58c6c-vv8cw\" (UID: \"8dd30f73-72f3-4733-b5c8-f96753750fb8\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.937068 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ba71fca-3e78-49e5-bfe5-b831568118ba-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.944291 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.950671 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-hghfw" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.951279 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ztq6g" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.976303 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-twxn6"] Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.977520 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.987101 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-268bs" Oct 02 07:31:41 crc kubenswrapper[4829]: I1002 07:31:41.993463 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-twxn6"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.015854 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.021356 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.026783 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-r92ns" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.029902 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.030522 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041009 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4x6b\" (UniqueName: \"kubernetes.io/projected/2ba71fca-3e78-49e5-bfe5-b831568118ba-kube-api-access-j4x6b\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041073 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj2hd\" (UniqueName: \"kubernetes.io/projected/c35efaf9-0b1b-4801-81c9-f8db24083049-kube-api-access-zj2hd\") pod \"octavia-operator-controller-manager-7b787867f4-hlq5c\" (UID: \"c35efaf9-0b1b-4801-81c9-f8db24083049\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041112 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc2dg\" (UniqueName: \"kubernetes.io/projected/fb49bfb5-3c2c-45eb-94d5-8061c919435d-kube-api-access-mc2dg\") pod \"swift-operator-controller-manager-84d6b4b759-mqflt\" (UID: \"fb49bfb5-3c2c-45eb-94d5-8061c919435d\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041147 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfbw5\" (UniqueName: \"kubernetes.io/projected/8dd30f73-72f3-4733-b5c8-f96753750fb8-kube-api-access-hfbw5\") pod \"placement-operator-controller-manager-589c58c6c-vv8cw\" (UID: \"8dd30f73-72f3-4733-b5c8-f96753750fb8\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041174 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ba71fca-3e78-49e5-bfe5-b831568118ba-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041205 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9bw\" (UniqueName: \"kubernetes.io/projected/71170543-3bd9-4d9f-9ad4-a5978a6f018b-kube-api-access-9s9bw\") pod \"telemetry-operator-controller-manager-b8d54b5d7-4wdcm\" (UID: \"71170543-3bd9-4d9f-9ad4-a5978a6f018b\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041261 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcwk5\" (UniqueName: \"kubernetes.io/projected/bf88c28f-c14a-4390-83ed-75dc3e41061d-kube-api-access-jcwk5\") pod \"nova-operator-controller-manager-64cd67b5cb-722dn\" (UID: \"bf88c28f-c14a-4390-83ed-75dc3e41061d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.041306 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdhrt\" (UniqueName: \"kubernetes.io/projected/3dee869a-23fe-4926-a561-67a87ca83103-kube-api-access-cdhrt\") pod \"ovn-operator-controller-manager-9976ff44c-5cqsw\" (UID: \"3dee869a-23fe-4926-a561-67a87ca83103\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:31:42 crc kubenswrapper[4829]: E1002 07:31:42.042011 4829 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 07:31:42 crc kubenswrapper[4829]: E1002 07:31:42.042053 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ba71fca-3e78-49e5-bfe5-b831568118ba-cert podName:2ba71fca-3e78-49e5-bfe5-b831568118ba nodeName:}" failed. No retries permitted until 2025-10-02 07:31:42.542038521 +0000 UTC m=+893.881686926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ba71fca-3e78-49e5-bfe5-b831568118ba-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-swsvz" (UID: "2ba71fca-3e78-49e5-bfe5-b831568118ba") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.045245 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.050509 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.078425 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcwk5\" (UniqueName: \"kubernetes.io/projected/bf88c28f-c14a-4390-83ed-75dc3e41061d-kube-api-access-jcwk5\") pod \"nova-operator-controller-manager-64cd67b5cb-722dn\" (UID: \"bf88c28f-c14a-4390-83ed-75dc3e41061d\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.079286 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdhrt\" (UniqueName: \"kubernetes.io/projected/3dee869a-23fe-4926-a561-67a87ca83103-kube-api-access-cdhrt\") pod \"ovn-operator-controller-manager-9976ff44c-5cqsw\" (UID: \"3dee869a-23fe-4926-a561-67a87ca83103\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.079895 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfbw5\" (UniqueName: \"kubernetes.io/projected/8dd30f73-72f3-4733-b5c8-f96753750fb8-kube-api-access-hfbw5\") pod \"placement-operator-controller-manager-589c58c6c-vv8cw\" (UID: \"8dd30f73-72f3-4733-b5c8-f96753750fb8\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.085383 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj2hd\" (UniqueName: \"kubernetes.io/projected/c35efaf9-0b1b-4801-81c9-f8db24083049-kube-api-access-zj2hd\") pod \"octavia-operator-controller-manager-7b787867f4-hlq5c\" (UID: \"c35efaf9-0b1b-4801-81c9-f8db24083049\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.092552 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.109514 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.138505 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4x6b\" (UniqueName: \"kubernetes.io/projected/2ba71fca-3e78-49e5-bfe5-b831568118ba-kube-api-access-j4x6b\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.165584 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc2dg\" (UniqueName: \"kubernetes.io/projected/fb49bfb5-3c2c-45eb-94d5-8061c919435d-kube-api-access-mc2dg\") pod \"swift-operator-controller-manager-84d6b4b759-mqflt\" (UID: \"fb49bfb5-3c2c-45eb-94d5-8061c919435d\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.165673 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p6nn\" (UniqueName: \"kubernetes.io/projected/a3ae11b8-1d62-41ba-a63b-4441d2f70709-kube-api-access-7p6nn\") pod \"watcher-operator-controller-manager-6bf8759486-hrl49\" (UID: \"a3ae11b8-1d62-41ba-a63b-4441d2f70709\") " pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.165714 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9bw\" (UniqueName: \"kubernetes.io/projected/71170543-3bd9-4d9f-9ad4-a5978a6f018b-kube-api-access-9s9bw\") pod \"telemetry-operator-controller-manager-b8d54b5d7-4wdcm\" (UID: \"71170543-3bd9-4d9f-9ad4-a5978a6f018b\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.165741 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z428q\" (UniqueName: \"kubernetes.io/projected/8129c7b9-335b-4bda-8516-94a818a57591-kube-api-access-z428q\") pod \"test-operator-controller-manager-85777745bb-twxn6\" (UID: \"8129c7b9-335b-4bda-8516-94a818a57591\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.209982 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc2dg\" (UniqueName: \"kubernetes.io/projected/fb49bfb5-3c2c-45eb-94d5-8061c919435d-kube-api-access-mc2dg\") pod \"swift-operator-controller-manager-84d6b4b759-mqflt\" (UID: \"fb49bfb5-3c2c-45eb-94d5-8061c919435d\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.210051 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.218135 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f874976-r72h2"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.221335 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.222588 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f874976-r72h2"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.224995 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-r9zfd" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.225291 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9bw\" (UniqueName: \"kubernetes.io/projected/71170543-3bd9-4d9f-9ad4-a5978a6f018b-kube-api-access-9s9bw\") pod \"telemetry-operator-controller-manager-b8d54b5d7-4wdcm\" (UID: \"71170543-3bd9-4d9f-9ad4-a5978a6f018b\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.229428 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.257555 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.261091 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.262039 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.263715 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6rrwm" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.267091 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p6nn\" (UniqueName: \"kubernetes.io/projected/a3ae11b8-1d62-41ba-a63b-4441d2f70709-kube-api-access-7p6nn\") pod \"watcher-operator-controller-manager-6bf8759486-hrl49\" (UID: \"a3ae11b8-1d62-41ba-a63b-4441d2f70709\") " pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.267132 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z428q\" (UniqueName: \"kubernetes.io/projected/8129c7b9-335b-4bda-8516-94a818a57591-kube-api-access-z428q\") pod \"test-operator-controller-manager-85777745bb-twxn6\" (UID: \"8129c7b9-335b-4bda-8516-94a818a57591\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.286276 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p6nn\" (UniqueName: \"kubernetes.io/projected/a3ae11b8-1d62-41ba-a63b-4441d2f70709-kube-api-access-7p6nn\") pod \"watcher-operator-controller-manager-6bf8759486-hrl49\" (UID: \"a3ae11b8-1d62-41ba-a63b-4441d2f70709\") " pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.294994 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z428q\" (UniqueName: \"kubernetes.io/projected/8129c7b9-335b-4bda-8516-94a818a57591-kube-api-access-z428q\") pod \"test-operator-controller-manager-85777745bb-twxn6\" (UID: \"8129c7b9-335b-4bda-8516-94a818a57591\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.296480 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.315016 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.349844 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.370528 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-cert\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.370590 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgpbk\" (UniqueName: \"kubernetes.io/projected/1c017022-5036-4312-b914-8aeb5439a9ef-kube-api-access-qgpbk\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-d4t94\" (UID: \"1c017022-5036-4312-b914-8aeb5439a9ef\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.370821 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vfg6\" (UniqueName: \"kubernetes.io/projected/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-kube-api-access-6vfg6\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.395698 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.406272 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.440257 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.471478 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-cert\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.471536 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgpbk\" (UniqueName: \"kubernetes.io/projected/1c017022-5036-4312-b914-8aeb5439a9ef-kube-api-access-qgpbk\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-d4t94\" (UID: \"1c017022-5036-4312-b914-8aeb5439a9ef\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.471569 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vfg6\" (UniqueName: \"kubernetes.io/projected/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-kube-api-access-6vfg6\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: E1002 07:31:42.471672 4829 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 07:31:42 crc kubenswrapper[4829]: E1002 07:31:42.471743 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-cert podName:f3a3eba9-d7d0-4796-acdc-2b9525a17e7a nodeName:}" failed. No retries permitted until 2025-10-02 07:31:42.971726155 +0000 UTC m=+894.311374560 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-cert") pod "openstack-operator-controller-manager-79f874976-r72h2" (UID: "f3a3eba9-d7d0-4796-acdc-2b9525a17e7a") : secret "webhook-server-cert" not found Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.501670 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.502775 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vfg6\" (UniqueName: \"kubernetes.io/projected/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-kube-api-access-6vfg6\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.512729 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgpbk\" (UniqueName: \"kubernetes.io/projected/1c017022-5036-4312-b914-8aeb5439a9ef-kube-api-access-qgpbk\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-d4t94\" (UID: \"1c017022-5036-4312-b914-8aeb5439a9ef\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.573618 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ba71fca-3e78-49e5-bfe5-b831568118ba-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.577862 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ba71fca-3e78-49e5-bfe5-b831568118ba-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-swsvz\" (UID: \"2ba71fca-3e78-49e5-bfe5-b831568118ba\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.590671 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.643679 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.752755 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.926159 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.939658 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.951514 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.959636 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.970291 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.976793 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs"] Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.981942 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-cert\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:42 crc kubenswrapper[4829]: I1002 07:31:42.986306 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3a3eba9-d7d0-4796-acdc-2b9525a17e7a-cert\") pod \"openstack-operator-controller-manager-79f874976-r72h2\" (UID: \"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a\") " pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.090437 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.098645 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn"] Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.101203 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3430c81b_20da_4871_9f8c_a408573677ec.slice/crio-9281eb330d74952ea8afca566b3ec1fc1b995eb33a8dc68690d6454429e40e85 WatchSource:0}: Error finding container 9281eb330d74952ea8afca566b3ec1fc1b995eb33a8dc68690d6454429e40e85: Status 404 returned error can't find the container with id 9281eb330d74952ea8afca566b3ec1fc1b995eb33a8dc68690d6454429e40e85 Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.102003 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc555f81b_a5f9_4b4b_9ae0_5253ebb12099.slice/crio-6de5a93f3df98767b0552ca07dde032ee3efd8cbeac54c3ac0a539eee9d5516b WatchSource:0}: Error finding container 6de5a93f3df98767b0552ca07dde032ee3efd8cbeac54c3ac0a539eee9d5516b: Status 404 returned error can't find the container with id 6de5a93f3df98767b0552ca07dde032ee3efd8cbeac54c3ac0a539eee9d5516b Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.109674 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8"] Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.116127 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5607af74_7a80_4b1c_9b33_690d3f7fe974.slice/crio-249fa1637802ed969bc953e9b64443cf8e55b62b199fd7dbf8d6bcdea8930702 WatchSource:0}: Error finding container 249fa1637802ed969bc953e9b64443cf8e55b62b199fd7dbf8d6bcdea8930702: Status 404 returned error can't find the container with id 249fa1637802ed969bc953e9b64443cf8e55b62b199fd7dbf8d6bcdea8930702 Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.221066 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" event={"ID":"6f685eab-a458-4058-9453-5c4a389758e0","Type":"ContainerStarted","Data":"1788fe478742bfe7459b0891c00a91a45658179c7a6794857c0bd2133efd3e46"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.223357 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" event={"ID":"3f4b21a4-b546-4f69-a432-05ba29b24085","Type":"ContainerStarted","Data":"0e6c1bde00647fe186a7ec7609d1656e8a44468efe3de1e0da2a5902877c404e"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.225724 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" event={"ID":"3430c81b-20da-4871-9f8c-a408573677ec","Type":"ContainerStarted","Data":"9281eb330d74952ea8afca566b3ec1fc1b995eb33a8dc68690d6454429e40e85"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.229604 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" event={"ID":"104543c5-16f7-47d1-b2cb-dedd729a87b7","Type":"ContainerStarted","Data":"2d5f1d5ba96f9a51d3e60dfaea128169486ae303f2905fff7f08d08e249ffaf2"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.234592 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.236442 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.237422 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" event={"ID":"895426fb-40c5-426b-ab48-f5ac31d22e06","Type":"ContainerStarted","Data":"b704bff6a3cea472870ab9a7fae8ea0a4f2bb4c0d43ae807e678dcb8554b7f1c"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.238533 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" event={"ID":"88cfec3f-d6c0-4007-a845-a3c4e56bcb63","Type":"ContainerStarted","Data":"b340e38555181499e1d0eca8d1515746859d7d497f70469fab3c079b22dda0ea"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.239376 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" event={"ID":"f588d058-d57b-4824-b59d-40731650907f","Type":"ContainerStarted","Data":"56ee4410f2f68971ab140802786b4284c0bbc7d6eff901b41d65be8478677aba"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.240784 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" event={"ID":"c555f81b-a5f9-4b4b-9ae0-5253ebb12099","Type":"ContainerStarted","Data":"6de5a93f3df98767b0552ca07dde032ee3efd8cbeac54c3ac0a539eee9d5516b"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.242480 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" event={"ID":"5607af74-7a80-4b1c-9b33-690d3f7fe974","Type":"ContainerStarted","Data":"249fa1637802ed969bc953e9b64443cf8e55b62b199fd7dbf8d6bcdea8930702"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.242509 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.249144 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" event={"ID":"349c96b5-e59f-4eac-8646-e5459775391c","Type":"ContainerStarted","Data":"441fa31fcbf8bc1fa555c8e74b90fca6da6b7a2faed327b988e5bd69efd705e8"} Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.251084 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71170543_3bd9_4d9f_9ad4_a5978a6f018b.slice/crio-f9229627c58f5334119cabff8441f4711dce69ab58e0ab2a68e1d7b947e563ac WatchSource:0}: Error finding container f9229627c58f5334119cabff8441f4711dce69ab58e0ab2a68e1d7b947e563ac: Status 404 returned error can't find the container with id f9229627c58f5334119cabff8441f4711dce69ab58e0ab2a68e1d7b947e563ac Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.251418 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.253413 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.254404 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" event={"ID":"e46a78e2-3b2a-47c5-8a9a-03877339340d","Type":"ContainerStarted","Data":"f8f48dd2fa35d9ab13cc9dc1afe659f196f89ae5948f11fa630adc5a93babd7e"} Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.256926 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49"] Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.263350 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dee869a_23fe_4926_a561_67a87ca83103.slice/crio-79c50770abd89687b4d60ca189d9fd02a234b9403920b9b198373506dda8bed5 WatchSource:0}: Error finding container 79c50770abd89687b4d60ca189d9fd02a234b9403920b9b198373506dda8bed5: Status 404 returned error can't find the container with id 79c50770abd89687b4d60ca189d9fd02a234b9403920b9b198373506dda8bed5 Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.265296 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn"] Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.271131 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8q6hr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-849d5b9b84-2rk2r_openstack-operators(d697f00d-25f4-48fb-808e-1a5fb90ff5fa): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.272085 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.278962 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt"] Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.281560 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hfbw5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-589c58c6c-vv8cw_openstack-operators(8dd30f73-72f3-4733-b5c8-f96753750fb8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.281647 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.65:5001/openstack-k8s-operators/watcher-operator:76aea9f1d102f74b05bfc24ed77fcfd8194edf1c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7p6nn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6bf8759486-hrl49_openstack-operators(a3ae11b8-1d62-41ba-a63b-4441d2f70709): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.288287 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf88c28f_c14a_4390_83ed_75dc3e41061d.slice/crio-b1858596f83e429e14c2fbeeca887146568fa1ac3ebb161d944936d7fa6eec82 WatchSource:0}: Error finding container b1858596f83e429e14c2fbeeca887146568fa1ac3ebb161d944936d7fa6eec82: Status 404 returned error can't find the container with id b1858596f83e429e14c2fbeeca887146568fa1ac3ebb161d944936d7fa6eec82 Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.290506 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jcwk5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-722dn_openstack-operators(bf88c28f-c14a-4390-83ed-75dc3e41061d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.296462 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zj2hd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7b787867f4-hlq5c_openstack-operators(c35efaf9-0b1b-4801-81c9-f8db24083049): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.298400 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mc2dg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-84d6b4b759-mqflt_openstack-operators(fb49bfb5-3c2c-45eb-94d5-8061c919435d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.440484 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-twxn6"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.444466 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94"] Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.464181 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz"] Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.477711 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qgpbk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-d4t94_openstack-operators(1c017022-5036-4312-b914-8aeb5439a9ef): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.479392 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" podUID="1c017022-5036-4312-b914-8aeb5439a9ef" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.498704 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j4x6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-swsvz_openstack-operators(2ba71fca-3e78-49e5-bfe5-b831568118ba): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.534687 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" podUID="d697f00d-25f4-48fb-808e-1a5fb90ff5fa" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.537489 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" podUID="bf88c28f-c14a-4390-83ed-75dc3e41061d" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.550446 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" podUID="c35efaf9-0b1b-4801-81c9-f8db24083049" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.562626 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" podUID="8dd30f73-72f3-4733-b5c8-f96753750fb8" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.566963 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" podUID="fb49bfb5-3c2c-45eb-94d5-8061c919435d" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.602781 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" podUID="a3ae11b8-1d62-41ba-a63b-4441d2f70709" Oct 02 07:31:43 crc kubenswrapper[4829]: E1002 07:31:43.666414 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" podUID="2ba71fca-3e78-49e5-bfe5-b831568118ba" Oct 02 07:31:43 crc kubenswrapper[4829]: I1002 07:31:43.755106 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79f874976-r72h2"] Oct 02 07:31:43 crc kubenswrapper[4829]: W1002 07:31:43.762569 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3a3eba9_d7d0_4796_acdc_2b9525a17e7a.slice/crio-f7a40543dc9b53f13b17f3878d9a856b73fe7c963dbfd969c7452c01052218c5 WatchSource:0}: Error finding container f7a40543dc9b53f13b17f3878d9a856b73fe7c963dbfd969c7452c01052218c5: Status 404 returned error can't find the container with id f7a40543dc9b53f13b17f3878d9a856b73fe7c963dbfd969c7452c01052218c5 Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.286991 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" event={"ID":"8129c7b9-335b-4bda-8516-94a818a57591","Type":"ContainerStarted","Data":"0771a02a4f01c650eee11e8dfbbc7af46270b7b9027d1a61017e8dee10069d44"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.290781 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" event={"ID":"d697f00d-25f4-48fb-808e-1a5fb90ff5fa","Type":"ContainerStarted","Data":"2c60caa429a7bf491030fd42ce0e181d836903ff5311ce6fb28382e5f6a88a71"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.290831 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" event={"ID":"d697f00d-25f4-48fb-808e-1a5fb90ff5fa","Type":"ContainerStarted","Data":"ceaee7ab0d0212eaf48db2ade5a5c1500b6476f1e87de5390ad5f06c0d0993f5"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.292629 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" event={"ID":"8dd30f73-72f3-4733-b5c8-f96753750fb8","Type":"ContainerStarted","Data":"88e00bfb1fbcd7278381c373ce1f15936bef640187a42ea478530745b8103b6a"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.292661 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" event={"ID":"8dd30f73-72f3-4733-b5c8-f96753750fb8","Type":"ContainerStarted","Data":"f3c0adf5eec152e953c63abba2ea5a9c7a340e504fd42d590f2ec92d05840f9e"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.292937 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" podUID="d697f00d-25f4-48fb-808e-1a5fb90ff5fa" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.297933 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" event={"ID":"bf88c28f-c14a-4390-83ed-75dc3e41061d","Type":"ContainerStarted","Data":"a589fb3246869352ad4c438244373bf55723cc8e76e05b0a87b6696d74772bf2"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.297999 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" event={"ID":"bf88c28f-c14a-4390-83ed-75dc3e41061d","Type":"ContainerStarted","Data":"b1858596f83e429e14c2fbeeca887146568fa1ac3ebb161d944936d7fa6eec82"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.299287 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" podUID="bf88c28f-c14a-4390-83ed-75dc3e41061d" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.305448 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" event={"ID":"fb49bfb5-3c2c-45eb-94d5-8061c919435d","Type":"ContainerStarted","Data":"30d328954a237f81c4097fb3e133872b0cfea96b1fb2faff48b05ab5b38d180a"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.305490 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" event={"ID":"fb49bfb5-3c2c-45eb-94d5-8061c919435d","Type":"ContainerStarted","Data":"81d25f5c0cdd3e1f118265ac9782f5a51ead29fbaf8285ca7384104590c1f9e1"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.306679 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" podUID="fb49bfb5-3c2c-45eb-94d5-8061c919435d" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.311290 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" event={"ID":"a3ae11b8-1d62-41ba-a63b-4441d2f70709","Type":"ContainerStarted","Data":"f2909c90fb2350c9a45e13875e40ddc2aa68020597fc1b90ed274d25822eefc9"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.311335 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" event={"ID":"a3ae11b8-1d62-41ba-a63b-4441d2f70709","Type":"ContainerStarted","Data":"31502fb8694f34ddfeb00046ea4aa91eefb76a0a047a001cff29767873e33cfb"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.312935 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.65:5001/openstack-k8s-operators/watcher-operator:76aea9f1d102f74b05bfc24ed77fcfd8194edf1c\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" podUID="a3ae11b8-1d62-41ba-a63b-4441d2f70709" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.313699 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" event={"ID":"71170543-3bd9-4d9f-9ad4-a5978a6f018b","Type":"ContainerStarted","Data":"f9229627c58f5334119cabff8441f4711dce69ab58e0ab2a68e1d7b947e563ac"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.317141 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" podUID="8dd30f73-72f3-4733-b5c8-f96753750fb8" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.317305 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" event={"ID":"3dee869a-23fe-4926-a561-67a87ca83103","Type":"ContainerStarted","Data":"79c50770abd89687b4d60ca189d9fd02a234b9403920b9b198373506dda8bed5"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.330455 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" event={"ID":"c35efaf9-0b1b-4801-81c9-f8db24083049","Type":"ContainerStarted","Data":"f6dc5fe7743555927417222d22c1f5ba39a6a35973c08d838619e7337fe05129"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.330498 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" event={"ID":"c35efaf9-0b1b-4801-81c9-f8db24083049","Type":"ContainerStarted","Data":"1887d5ebcfcf6a131528f14e9dad1b0e53ca220a9798195be67c0e4086e4bfc7"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.333855 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" podUID="c35efaf9-0b1b-4801-81c9-f8db24083049" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.334720 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" event={"ID":"1c017022-5036-4312-b914-8aeb5439a9ef","Type":"ContainerStarted","Data":"fe24f7f5bd3a9d76c10133b74fb97a50f1049b30207448ef80cd81a7b76ea84a"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.339835 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" podUID="1c017022-5036-4312-b914-8aeb5439a9ef" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.341110 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" event={"ID":"2ba71fca-3e78-49e5-bfe5-b831568118ba","Type":"ContainerStarted","Data":"975ca8f9add61717ba0257287f330c9d1f53d5659e0e67749fa1fc105a62b096"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.341147 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" event={"ID":"2ba71fca-3e78-49e5-bfe5-b831568118ba","Type":"ContainerStarted","Data":"2985f17a19b82efb2c34d7a996a1de21133ee052ae829e0cdb05dc2a0b0627fb"} Oct 02 07:31:44 crc kubenswrapper[4829]: E1002 07:31:44.343824 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" podUID="2ba71fca-3e78-49e5-bfe5-b831568118ba" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.347060 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" event={"ID":"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a","Type":"ContainerStarted","Data":"704355d283093b18cebc03d18873ce02de03ff2f348f39b30a02b77ac67533e3"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.347093 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" event={"ID":"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a","Type":"ContainerStarted","Data":"88ba55e44d3a2f3f9cb6541af445cb2209977738fd4c1cc646f9607c04559f9a"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.347103 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" event={"ID":"f3a3eba9-d7d0-4796-acdc-2b9525a17e7a","Type":"ContainerStarted","Data":"f7a40543dc9b53f13b17f3878d9a856b73fe7c963dbfd969c7452c01052218c5"} Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.347706 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:44 crc kubenswrapper[4829]: I1002 07:31:44.481137 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" podStartSLOduration=2.481116411 podStartE2EDuration="2.481116411s" podCreationTimestamp="2025-10-02 07:31:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:31:44.470313454 +0000 UTC m=+895.809961859" watchObservedRunningTime="2025-10-02 07:31:44.481116411 +0000 UTC m=+895.820764816" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.355984 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" podUID="fb49bfb5-3c2c-45eb-94d5-8061c919435d" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.357643 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" podUID="c35efaf9-0b1b-4801-81c9-f8db24083049" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.357651 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:acdeebaa51f962066f42f38b6c2d34a62fc6a24f58f9ee63d61b1e0cafbb29f8\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" podUID="d697f00d-25f4-48fb-808e-1a5fb90ff5fa" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.359025 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a6b3408d79df6b6d4a467e49defaa4a9d9c088c94d0605a4fee0030c9ccc84d2\\\"\"" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" podUID="8dd30f73-72f3-4733-b5c8-f96753750fb8" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.359097 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" podUID="1c017022-5036-4312-b914-8aeb5439a9ef" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.359142 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" podUID="2ba71fca-3e78-49e5-bfe5-b831568118ba" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.359212 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.65:5001/openstack-k8s-operators/watcher-operator:76aea9f1d102f74b05bfc24ed77fcfd8194edf1c\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" podUID="a3ae11b8-1d62-41ba-a63b-4441d2f70709" Oct 02 07:31:45 crc kubenswrapper[4829]: E1002 07:31:45.360059 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" podUID="bf88c28f-c14a-4390-83ed-75dc3e41061d" Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.426460 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" event={"ID":"3f4b21a4-b546-4f69-a432-05ba29b24085","Type":"ContainerStarted","Data":"9cb203e973e64c9d841757042a2f1a931a2225bb34a49fbb6813ca06abc0a0d7"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.450980 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" event={"ID":"8129c7b9-335b-4bda-8516-94a818a57591","Type":"ContainerStarted","Data":"473be96d4ab7437f708bdd6ed2dedece3ea29ced4705d08a098e4695c2c32a44"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.453445 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" event={"ID":"6f685eab-a458-4058-9453-5c4a389758e0","Type":"ContainerStarted","Data":"d21fd94742dca7d4f8e5b12ed811641a5c119b63d92885c3b90e7610a58beb62"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.470771 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" event={"ID":"5607af74-7a80-4b1c-9b33-690d3f7fe974","Type":"ContainerStarted","Data":"a315bcf076a77ebdf85f3d0428bf1e20fa4dfd8ca87a1acfc1b092a51eaa9388"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.475241 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" event={"ID":"895426fb-40c5-426b-ab48-f5ac31d22e06","Type":"ContainerStarted","Data":"4756c19b4a9bf9d968fd7e16ec5a10d86e23dbacc01f8c61e9b96f4c65541411"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.489469 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" event={"ID":"3dee869a-23fe-4926-a561-67a87ca83103","Type":"ContainerStarted","Data":"e798a488eb12846425f455981d5644e5cb1f6cfdb29fbcb1a89763f7956a873f"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.501772 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" event={"ID":"349c96b5-e59f-4eac-8646-e5459775391c","Type":"ContainerStarted","Data":"3097093d7fa897f6ff6a21f2e9cd034a2d2fb3158f454273434acba075a44650"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.520833 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" event={"ID":"f588d058-d57b-4824-b59d-40731650907f","Type":"ContainerStarted","Data":"72f28fe523245344848711eccc67fd743e88d2d55c03287bfe28928b955ef20b"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.522399 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" event={"ID":"c555f81b-a5f9-4b4b-9ae0-5253ebb12099","Type":"ContainerStarted","Data":"3ba82e9a0186a61cc241e2c9961f5000c69b098f6a6691ea30b9286e992ad1d8"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.522427 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" event={"ID":"c555f81b-a5f9-4b4b-9ae0-5253ebb12099","Type":"ContainerStarted","Data":"1a42b1d923865613682defda6cde8bdc4c1f46abc2d5cd612bb93947df0a369a"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.522786 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.535716 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" event={"ID":"3430c81b-20da-4871-9f8c-a408573677ec","Type":"ContainerStarted","Data":"5ecbc6c56302a0337960000fcfc8f758fc7c9efe2ef0f0024811b1fd5abd9de6"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.548454 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" event={"ID":"104543c5-16f7-47d1-b2cb-dedd729a87b7","Type":"ContainerStarted","Data":"80237ef3c4da51fcd676667f7d70b32a148b6fd3d288ca58dc8ba190c3367df3"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.572797 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" podStartSLOduration=3.24172224 podStartE2EDuration="11.572780699s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.106441757 +0000 UTC m=+894.446090202" lastFinishedPulling="2025-10-02 07:31:51.437500256 +0000 UTC m=+902.777148661" observedRunningTime="2025-10-02 07:31:52.558580753 +0000 UTC m=+903.898229178" watchObservedRunningTime="2025-10-02 07:31:52.572780699 +0000 UTC m=+903.912429104" Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.578586 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" event={"ID":"71170543-3bd9-4d9f-9ad4-a5978a6f018b","Type":"ContainerStarted","Data":"fd318e27a6ca2543a94731c794f95e91a454c65cd636fc0edcd6c95b33bb8a57"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.583091 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" event={"ID":"e46a78e2-3b2a-47c5-8a9a-03877339340d","Type":"ContainerStarted","Data":"908de72e0a611cf66dac30f1cee7eacdffac3b659f6b8a0c96c5834e1a2c3886"} Oct 02 07:31:52 crc kubenswrapper[4829]: I1002 07:31:52.584267 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" event={"ID":"88cfec3f-d6c0-4007-a845-a3c4e56bcb63","Type":"ContainerStarted","Data":"d4a0bfbdb47be4f1fdde2058f8fa18b1e741ba396f3a86a2d9722ae6056f5a0d"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.240583 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-79f874976-r72h2" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.592608 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" event={"ID":"88cfec3f-d6c0-4007-a845-a3c4e56bcb63","Type":"ContainerStarted","Data":"c7694d90abf781e20d3b8841877235bac5dffead81b611a703013a12f9707648"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.593668 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.595965 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" event={"ID":"e46a78e2-3b2a-47c5-8a9a-03877339340d","Type":"ContainerStarted","Data":"7f9504385be83115b926e752f9c04220a6f8f5ad098e2ccbecb427110a06a8ce"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.596101 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.601561 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" event={"ID":"6f685eab-a458-4058-9453-5c4a389758e0","Type":"ContainerStarted","Data":"284f09dc9414a4a2739ce9bf311ca82f56178c7418a7e026851c215f5cce289a"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.603146 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" event={"ID":"8129c7b9-335b-4bda-8516-94a818a57591","Type":"ContainerStarted","Data":"b6d069d9a7b2646881be24cdea42042f3e0f813cce7389f6e9d522475ea439c1"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.603316 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.605323 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" event={"ID":"5607af74-7a80-4b1c-9b33-690d3f7fe974","Type":"ContainerStarted","Data":"48a3660b0f8eff5cfec39bc0dd50f8225ca160678ef336eec2240ea7649f4628"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.605429 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.607009 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" event={"ID":"71170543-3bd9-4d9f-9ad4-a5978a6f018b","Type":"ContainerStarted","Data":"34836afae50aacc1fc8402de303cfc88a98ae21a0a9d560e3e07dc74a92a27fa"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.607517 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.609687 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" event={"ID":"3dee869a-23fe-4926-a561-67a87ca83103","Type":"ContainerStarted","Data":"71a12aeb1fde61a46c24c2b5194a53f04ac7405cf73cfdbdb24fcc54cb49e655"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.610208 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.612591 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" event={"ID":"895426fb-40c5-426b-ab48-f5ac31d22e06","Type":"ContainerStarted","Data":"8e73db4e7e694b0ae7413ac74d72f75cc09f618a367ff0259db75c7d1798dc5c"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.612693 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" podStartSLOduration=4.14441101 podStartE2EDuration="12.612681267s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.96304383 +0000 UTC m=+894.302692235" lastFinishedPulling="2025-10-02 07:31:51.431314087 +0000 UTC m=+902.770962492" observedRunningTime="2025-10-02 07:31:53.610519848 +0000 UTC m=+904.950168273" watchObservedRunningTime="2025-10-02 07:31:53.612681267 +0000 UTC m=+904.952329672" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.612962 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.618600 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" event={"ID":"3430c81b-20da-4871-9f8c-a408573677ec","Type":"ContainerStarted","Data":"3dc2366986c9eb33abb3eda5279b6e3bfe51259290d00dc2b062a2de308eedb0"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.618711 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.621270 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" event={"ID":"3f4b21a4-b546-4f69-a432-05ba29b24085","Type":"ContainerStarted","Data":"55dc611ca76a05a0b0770b6a03aa9f1c709ea82ce0424cc741737630d35063ae"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.621852 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.625824 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" event={"ID":"104543c5-16f7-47d1-b2cb-dedd729a87b7","Type":"ContainerStarted","Data":"741f4a0313c6b808e57d65c4e9b484b81d9f3356bcc21393d0dcc38c329e7523"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.625960 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.631977 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" event={"ID":"349c96b5-e59f-4eac-8646-e5459775391c","Type":"ContainerStarted","Data":"b425ba0a148b5f583aeddd4ff3cf5d9c1bc2ebb4985174c7d6809c2a460910f8"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.632200 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.633017 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" podStartSLOduration=4.477855063 podStartE2EDuration="12.63300527s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.264559847 +0000 UTC m=+894.604208262" lastFinishedPulling="2025-10-02 07:31:51.419710064 +0000 UTC m=+902.759358469" observedRunningTime="2025-10-02 07:31:53.630615653 +0000 UTC m=+904.970264068" watchObservedRunningTime="2025-10-02 07:31:53.63300527 +0000 UTC m=+904.972653685" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.635614 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" event={"ID":"f588d058-d57b-4824-b59d-40731650907f","Type":"ContainerStarted","Data":"44c7c9c4879682dfe47b32649e57bdff3c29020657eb832b021c650e28320619"} Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.635761 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.648949 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" podStartSLOduration=4.476524329 podStartE2EDuration="12.648932351s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.257853121 +0000 UTC m=+894.597501526" lastFinishedPulling="2025-10-02 07:31:51.430261143 +0000 UTC m=+902.769909548" observedRunningTime="2025-10-02 07:31:53.646155602 +0000 UTC m=+904.985804007" watchObservedRunningTime="2025-10-02 07:31:53.648932351 +0000 UTC m=+904.988580756" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.669511 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" podStartSLOduration=4.218993866 podStartE2EDuration="12.669489542s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.98980332 +0000 UTC m=+894.329451725" lastFinishedPulling="2025-10-02 07:31:51.440298986 +0000 UTC m=+902.779947401" observedRunningTime="2025-10-02 07:31:53.663141369 +0000 UTC m=+905.002789794" watchObservedRunningTime="2025-10-02 07:31:53.669489542 +0000 UTC m=+905.009137957" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.690523 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" podStartSLOduration=4.71465917 podStartE2EDuration="12.690501807s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.462820396 +0000 UTC m=+894.802468801" lastFinishedPulling="2025-10-02 07:31:51.438663033 +0000 UTC m=+902.778311438" observedRunningTime="2025-10-02 07:31:53.682998196 +0000 UTC m=+905.022646601" watchObservedRunningTime="2025-10-02 07:31:53.690501807 +0000 UTC m=+905.030150212" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.712404 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" podStartSLOduration=4.413710522 podStartE2EDuration="12.71238307s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.121920705 +0000 UTC m=+894.461569130" lastFinishedPulling="2025-10-02 07:31:51.420593273 +0000 UTC m=+902.760241678" observedRunningTime="2025-10-02 07:31:53.710400117 +0000 UTC m=+905.050048532" watchObservedRunningTime="2025-10-02 07:31:53.71238307 +0000 UTC m=+905.052031475" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.724907 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" podStartSLOduration=4.251948825 podStartE2EDuration="12.724890372s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.962851884 +0000 UTC m=+894.302500289" lastFinishedPulling="2025-10-02 07:31:51.435793431 +0000 UTC m=+902.775441836" observedRunningTime="2025-10-02 07:31:53.723508768 +0000 UTC m=+905.063157173" watchObservedRunningTime="2025-10-02 07:31:53.724890372 +0000 UTC m=+905.064538777" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.746484 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" podStartSLOduration=3.818112968 podStartE2EDuration="12.746470096s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.501905735 +0000 UTC m=+893.841554140" lastFinishedPulling="2025-10-02 07:31:51.430262863 +0000 UTC m=+902.769911268" observedRunningTime="2025-10-02 07:31:53.742352452 +0000 UTC m=+905.082000857" watchObservedRunningTime="2025-10-02 07:31:53.746470096 +0000 UTC m=+905.086118501" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.761509 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" podStartSLOduration=4.452215219 podStartE2EDuration="12.761488518s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.110474267 +0000 UTC m=+894.450122672" lastFinishedPulling="2025-10-02 07:31:51.419747566 +0000 UTC m=+902.759395971" observedRunningTime="2025-10-02 07:31:53.758372638 +0000 UTC m=+905.098021063" watchObservedRunningTime="2025-10-02 07:31:53.761488518 +0000 UTC m=+905.101136913" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.777994 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" podStartSLOduration=4.329265757 podStartE2EDuration="12.777977977s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.990392298 +0000 UTC m=+894.330040703" lastFinishedPulling="2025-10-02 07:31:51.439104478 +0000 UTC m=+902.778752923" observedRunningTime="2025-10-02 07:31:53.771468818 +0000 UTC m=+905.111117223" watchObservedRunningTime="2025-10-02 07:31:53.777977977 +0000 UTC m=+905.117626382" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.790575 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" podStartSLOduration=4.282886769 podStartE2EDuration="12.790557252s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.963088721 +0000 UTC m=+894.302737126" lastFinishedPulling="2025-10-02 07:31:51.470759204 +0000 UTC m=+902.810407609" observedRunningTime="2025-10-02 07:31:53.788573168 +0000 UTC m=+905.128221583" watchObservedRunningTime="2025-10-02 07:31:53.790557252 +0000 UTC m=+905.130205657" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.812133 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" podStartSLOduration=3.917593323 podStartE2EDuration="12.812115484s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.501393909 +0000 UTC m=+893.841042314" lastFinishedPulling="2025-10-02 07:31:51.39591607 +0000 UTC m=+902.735564475" observedRunningTime="2025-10-02 07:31:53.806354619 +0000 UTC m=+905.146003034" watchObservedRunningTime="2025-10-02 07:31:53.812115484 +0000 UTC m=+905.151763889" Oct 02 07:31:53 crc kubenswrapper[4829]: I1002 07:31:53.830173 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" podStartSLOduration=4.366104343 podStartE2EDuration="12.830148694s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:42.941843029 +0000 UTC m=+894.281491434" lastFinishedPulling="2025-10-02 07:31:51.40588738 +0000 UTC m=+902.745535785" observedRunningTime="2025-10-02 07:31:53.824664087 +0000 UTC m=+905.164312502" watchObservedRunningTime="2025-10-02 07:31:53.830148694 +0000 UTC m=+905.169797099" Oct 02 07:31:54 crc kubenswrapper[4829]: I1002 07:31:54.649096 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:31:55 crc kubenswrapper[4829]: I1002 07:31:55.329661 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:31:55 crc kubenswrapper[4829]: I1002 07:31:55.329750 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:31:58 crc kubenswrapper[4829]: I1002 07:31:58.695703 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" event={"ID":"8dd30f73-72f3-4733-b5c8-f96753750fb8","Type":"ContainerStarted","Data":"349629bc600577ee0fa92bee3273551bd6af9d0c5a237d00082000b55670566b"} Oct 02 07:31:58 crc kubenswrapper[4829]: I1002 07:31:58.697186 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:31:58 crc kubenswrapper[4829]: I1002 07:31:58.717756 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" podStartSLOduration=3.179695475 podStartE2EDuration="17.717741123s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.281410238 +0000 UTC m=+894.621058643" lastFinishedPulling="2025-10-02 07:31:57.819455886 +0000 UTC m=+909.159104291" observedRunningTime="2025-10-02 07:31:58.714806709 +0000 UTC m=+910.054455144" watchObservedRunningTime="2025-10-02 07:31:58.717741123 +0000 UTC m=+910.057389528" Oct 02 07:32:00 crc kubenswrapper[4829]: I1002 07:32:00.720547 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" event={"ID":"bf88c28f-c14a-4390-83ed-75dc3e41061d","Type":"ContainerStarted","Data":"b078b121561a12024a1f8779d74293abe59400d49fffe39e1be55dabd0dc6c60"} Oct 02 07:32:00 crc kubenswrapper[4829]: I1002 07:32:00.721092 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:32:00 crc kubenswrapper[4829]: I1002 07:32:00.724622 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" event={"ID":"a3ae11b8-1d62-41ba-a63b-4441d2f70709","Type":"ContainerStarted","Data":"fa9ec2e3816924446fb279dc541b3b5678df5049fda719130f69aaf028387c7c"} Oct 02 07:32:00 crc kubenswrapper[4829]: I1002 07:32:00.724817 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:32:00 crc kubenswrapper[4829]: I1002 07:32:00.736292 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" podStartSLOduration=3.1063012309999998 podStartE2EDuration="19.736273469s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.290348515 +0000 UTC m=+894.629996940" lastFinishedPulling="2025-10-02 07:31:59.920320773 +0000 UTC m=+911.259969178" observedRunningTime="2025-10-02 07:32:00.733864232 +0000 UTC m=+912.073512637" watchObservedRunningTime="2025-10-02 07:32:00.736273469 +0000 UTC m=+912.075921874" Oct 02 07:32:00 crc kubenswrapper[4829]: I1002 07:32:00.756249 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" podStartSLOduration=3.123625234 podStartE2EDuration="19.756217674s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.281545583 +0000 UTC m=+894.621193988" lastFinishedPulling="2025-10-02 07:31:59.914138023 +0000 UTC m=+911.253786428" observedRunningTime="2025-10-02 07:32:00.749355262 +0000 UTC m=+912.089003667" watchObservedRunningTime="2025-10-02 07:32:00.756217674 +0000 UTC m=+912.095866079" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.655960 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-7wrzt" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.692817 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-nf7gj" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.697735 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-rs8gs" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.718317 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-5r5zb" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.740939 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-ztngc" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.750685 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" event={"ID":"1c017022-5036-4312-b914-8aeb5439a9ef","Type":"ContainerStarted","Data":"58c9b6f8fdcb8f09ca92379e7021d2a7dab825f3ef6c0aedb2bd5419cd07f5ea"} Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.758414 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-zc9kp" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.775459 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-cw4k8" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.799118 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-2njwr" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.811341 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-d4t94" podStartSLOduration=2.7155763029999997 podStartE2EDuration="19.811320528s" podCreationTimestamp="2025-10-02 07:31:42 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.477492517 +0000 UTC m=+894.817140922" lastFinishedPulling="2025-10-02 07:32:00.573236732 +0000 UTC m=+911.912885147" observedRunningTime="2025-10-02 07:32:01.804345413 +0000 UTC m=+913.143993828" watchObservedRunningTime="2025-10-02 07:32:01.811320528 +0000 UTC m=+913.150968933" Oct 02 07:32:01 crc kubenswrapper[4829]: I1002 07:32:01.848695 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-pf5r7" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.033559 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-tmszn" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.033857 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-p8wc8" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.213139 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-5cqsw" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.259445 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-vv8cw" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.443852 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-4wdcm" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.597604 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-twxn6" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.763277 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" event={"ID":"c35efaf9-0b1b-4801-81c9-f8db24083049","Type":"ContainerStarted","Data":"33be70fc6912dacd7334c785e9b4be2488a7c7c4b1f678efcfe867374cbd707c"} Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.764311 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.767302 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" event={"ID":"2ba71fca-3e78-49e5-bfe5-b831568118ba","Type":"ContainerStarted","Data":"7fabb8ee058165e94b357e4a97b3f729952fa6d6c91c3b2450d1c5c7a4ef8180"} Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.767798 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.783190 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" podStartSLOduration=3.434814346 podStartE2EDuration="21.783174743s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.296313687 +0000 UTC m=+894.635962092" lastFinishedPulling="2025-10-02 07:32:01.644674054 +0000 UTC m=+912.984322489" observedRunningTime="2025-10-02 07:32:02.776871859 +0000 UTC m=+914.116520274" watchObservedRunningTime="2025-10-02 07:32:02.783174743 +0000 UTC m=+914.122823158" Oct 02 07:32:02 crc kubenswrapper[4829]: I1002 07:32:02.809847 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" podStartSLOduration=4.219840455 podStartE2EDuration="21.809824924s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.495678321 +0000 UTC m=+894.835326726" lastFinishedPulling="2025-10-02 07:32:01.08566279 +0000 UTC m=+912.425311195" observedRunningTime="2025-10-02 07:32:02.804796631 +0000 UTC m=+914.144445036" watchObservedRunningTime="2025-10-02 07:32:02.809824924 +0000 UTC m=+914.149473329" Oct 02 07:32:03 crc kubenswrapper[4829]: I1002 07:32:03.777513 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" event={"ID":"d697f00d-25f4-48fb-808e-1a5fb90ff5fa","Type":"ContainerStarted","Data":"6118b9fc3dfa04425e3dbb69f0b44029c3f046cbc0158f95ce0dcc7bb740403c"} Oct 02 07:32:03 crc kubenswrapper[4829]: I1002 07:32:03.778044 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:32:03 crc kubenswrapper[4829]: I1002 07:32:03.779727 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" event={"ID":"fb49bfb5-3c2c-45eb-94d5-8061c919435d","Type":"ContainerStarted","Data":"29fda2ce050c3ee8e6041e7feb92e4eb1a93f26e043370240a3dad14775cadd8"} Oct 02 07:32:03 crc kubenswrapper[4829]: I1002 07:32:03.802926 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" podStartSLOduration=3.454952781 podStartE2EDuration="22.802904254s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.271033974 +0000 UTC m=+894.610682379" lastFinishedPulling="2025-10-02 07:32:02.618985447 +0000 UTC m=+913.958633852" observedRunningTime="2025-10-02 07:32:03.798397819 +0000 UTC m=+915.138046234" watchObservedRunningTime="2025-10-02 07:32:03.802904254 +0000 UTC m=+915.142552699" Oct 02 07:32:03 crc kubenswrapper[4829]: I1002 07:32:03.817171 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" podStartSLOduration=2.9014143900000002 podStartE2EDuration="22.817145505s" podCreationTimestamp="2025-10-02 07:31:41 +0000 UTC" firstStartedPulling="2025-10-02 07:31:43.298300611 +0000 UTC m=+894.637949016" lastFinishedPulling="2025-10-02 07:32:03.214031726 +0000 UTC m=+914.553680131" observedRunningTime="2025-10-02 07:32:03.815028966 +0000 UTC m=+915.154677371" watchObservedRunningTime="2025-10-02 07:32:03.817145505 +0000 UTC m=+915.156793930" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.054256 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2rk2r" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.097118 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-722dn" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.116541 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-hlq5c" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.317898 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6bf8759486-hrl49" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.350694 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.352466 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-mqflt" Oct 02 07:32:12 crc kubenswrapper[4829]: I1002 07:32:12.764552 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-swsvz" Oct 02 07:32:25 crc kubenswrapper[4829]: I1002 07:32:25.329342 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:32:25 crc kubenswrapper[4829]: I1002 07:32:25.329865 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.654284 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-f2m9b"] Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.656846 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.659981 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.660187 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.660355 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.660490 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-shqnv" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.666066 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-f2m9b"] Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.707998 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ljpcp"] Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.713353 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.717850 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.720993 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ljpcp"] Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.825994 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-config\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.826043 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvjs\" (UniqueName: \"kubernetes.io/projected/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-kube-api-access-qsvjs\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.826077 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trskc\" (UniqueName: \"kubernetes.io/projected/859cd9a2-b21a-4dd7-8649-97a677147c1d-kube-api-access-trskc\") pod \"dnsmasq-dns-675f4bcbfc-f2m9b\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.826121 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.826149 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859cd9a2-b21a-4dd7-8649-97a677147c1d-config\") pod \"dnsmasq-dns-675f4bcbfc-f2m9b\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.929968 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859cd9a2-b21a-4dd7-8649-97a677147c1d-config\") pod \"dnsmasq-dns-675f4bcbfc-f2m9b\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.928022 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859cd9a2-b21a-4dd7-8649-97a677147c1d-config\") pod \"dnsmasq-dns-675f4bcbfc-f2m9b\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.930222 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-config\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.930343 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvjs\" (UniqueName: \"kubernetes.io/projected/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-kube-api-access-qsvjs\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.930411 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trskc\" (UniqueName: \"kubernetes.io/projected/859cd9a2-b21a-4dd7-8649-97a677147c1d-kube-api-access-trskc\") pod \"dnsmasq-dns-675f4bcbfc-f2m9b\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.930504 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.931437 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.931664 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-config\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.957000 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvjs\" (UniqueName: \"kubernetes.io/projected/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-kube-api-access-qsvjs\") pod \"dnsmasq-dns-78dd6ddcc-ljpcp\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.965030 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trskc\" (UniqueName: \"kubernetes.io/projected/859cd9a2-b21a-4dd7-8649-97a677147c1d-kube-api-access-trskc\") pod \"dnsmasq-dns-675f4bcbfc-f2m9b\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:36 crc kubenswrapper[4829]: I1002 07:32:36.981287 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:37 crc kubenswrapper[4829]: I1002 07:32:37.030318 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:37 crc kubenswrapper[4829]: I1002 07:32:37.505930 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-f2m9b"] Oct 02 07:32:37 crc kubenswrapper[4829]: I1002 07:32:37.573300 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ljpcp"] Oct 02 07:32:37 crc kubenswrapper[4829]: W1002 07:32:37.584886 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1ecf4ff_3aeb_4c5e_a4e3_095dc8c663b3.slice/crio-044ac3cec8547a769a8ffac3e4eb2de4756f8b504cbc452467dc29ccf0a22c3d WatchSource:0}: Error finding container 044ac3cec8547a769a8ffac3e4eb2de4756f8b504cbc452467dc29ccf0a22c3d: Status 404 returned error can't find the container with id 044ac3cec8547a769a8ffac3e4eb2de4756f8b504cbc452467dc29ccf0a22c3d Oct 02 07:32:38 crc kubenswrapper[4829]: I1002 07:32:38.089025 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" event={"ID":"859cd9a2-b21a-4dd7-8649-97a677147c1d","Type":"ContainerStarted","Data":"07164f20696e4eaa7624c41985e0351ed9ff2310873d8a02fb0aa0f00d6a239d"} Oct 02 07:32:38 crc kubenswrapper[4829]: I1002 07:32:38.102390 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" event={"ID":"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3","Type":"ContainerStarted","Data":"044ac3cec8547a769a8ffac3e4eb2de4756f8b504cbc452467dc29ccf0a22c3d"} Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.582990 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-f2m9b"] Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.620257 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-mpllb"] Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.621517 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.630015 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-mpllb"] Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.686847 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.686942 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-config\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.686984 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzgc6\" (UniqueName: \"kubernetes.io/projected/bab06794-e96a-49ea-b1c7-d99ae28c3630-kube-api-access-mzgc6\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.788466 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.788533 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-config\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.788564 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzgc6\" (UniqueName: \"kubernetes.io/projected/bab06794-e96a-49ea-b1c7-d99ae28c3630-kube-api-access-mzgc6\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.789998 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-config\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.790032 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.832476 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzgc6\" (UniqueName: \"kubernetes.io/projected/bab06794-e96a-49ea-b1c7-d99ae28c3630-kube-api-access-mzgc6\") pod \"dnsmasq-dns-5ccc8479f9-mpllb\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.876571 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ljpcp"] Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.897711 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fjdw2"] Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.898915 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.915353 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fjdw2"] Oct 02 07:32:39 crc kubenswrapper[4829]: I1002 07:32:39.943511 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.096114 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmq5n\" (UniqueName: \"kubernetes.io/projected/9c93fa9a-4673-46f4-b432-ff794f518510-kube-api-access-jmq5n\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.096194 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.096730 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-config\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.198114 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmq5n\" (UniqueName: \"kubernetes.io/projected/9c93fa9a-4673-46f4-b432-ff794f518510-kube-api-access-jmq5n\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.198175 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.198247 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-config\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.199005 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-config\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.199782 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.216350 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmq5n\" (UniqueName: \"kubernetes.io/projected/9c93fa9a-4673-46f4-b432-ff794f518510-kube-api-access-jmq5n\") pod \"dnsmasq-dns-57d769cc4f-fjdw2\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.222178 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.743184 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.744583 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.746973 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.747114 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.747027 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.747432 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.747625 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.747667 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tdd4x" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.752477 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.757436 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.907818 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.907885 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.907957 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bf3dfd-bae4-4af6-a338-fad9857bfd29-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.907982 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908049 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908080 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908134 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908207 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908255 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7cbl\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-kube-api-access-x7cbl\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908366 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.908485 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bf3dfd-bae4-4af6-a338-fad9857bfd29-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:40 crc kubenswrapper[4829]: I1002 07:32:40.995118 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.000769 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.003007 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.004075 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sqk68" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.005041 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.005358 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.005440 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.005460 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.005858 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.006101 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.013734 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.013775 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.013850 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bf3dfd-bae4-4af6-a338-fad9857bfd29-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.013870 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.014511 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.014696 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.015139 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.015201 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.015204 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.015275 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017185 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017272 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017311 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017338 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7cbl\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-kube-api-access-x7cbl\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017411 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017933 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bf3dfd-bae4-4af6-a338-fad9857bfd29-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.017878 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.020001 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bf3dfd-bae4-4af6-a338-fad9857bfd29-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.026899 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.026927 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.030285 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bf3dfd-bae4-4af6-a338-fad9857bfd29-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.038127 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7cbl\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-kube-api-access-x7cbl\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.052953 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.074745 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119529 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119580 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119607 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs9k4\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-kube-api-access-rs9k4\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119647 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119791 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119847 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.119871 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-config-data\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.120009 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.120090 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.120146 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.120167 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222288 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222373 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222479 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222504 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222556 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs9k4\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-kube-api-access-rs9k4\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222581 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222653 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222669 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222696 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-config-data\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222794 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222794 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.222863 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.223019 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.223176 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.224128 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-config-data\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.224253 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.224284 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.225933 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.227726 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.227732 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.236611 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.237218 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs9k4\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-kube-api-access-rs9k4\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.244963 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " pod="openstack/rabbitmq-server-0" Oct 02 07:32:41 crc kubenswrapper[4829]: I1002 07:32:41.401444 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.631007 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.632589 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.634344 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.635859 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.635887 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.636208 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.639346 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-gmdzt" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.640865 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.644807 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779243 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779288 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779316 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-config-data-default\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779347 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779375 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-secrets\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779400 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf2zz\" (UniqueName: \"kubernetes.io/projected/0195e48a-f8c9-433b-a9a8-c3da38457280-kube-api-access-vf2zz\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779423 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-kolla-config\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779438 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0195e48a-f8c9-433b-a9a8-c3da38457280-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.779459 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880355 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880648 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-secrets\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880699 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf2zz\" (UniqueName: \"kubernetes.io/projected/0195e48a-f8c9-433b-a9a8-c3da38457280-kube-api-access-vf2zz\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880724 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-kolla-config\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880740 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0195e48a-f8c9-433b-a9a8-c3da38457280-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880762 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.880925 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.881008 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.881059 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-config-data-default\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.881376 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0195e48a-f8c9-433b-a9a8-c3da38457280-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.881633 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-kolla-config\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.882904 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-config-data-default\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.883488 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.884981 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0195e48a-f8c9-433b-a9a8-c3da38457280-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.887718 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-secrets\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.889328 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.906206 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0195e48a-f8c9-433b-a9a8-c3da38457280-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.909889 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.910586 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf2zz\" (UniqueName: \"kubernetes.io/projected/0195e48a-f8c9-433b-a9a8-c3da38457280-kube-api-access-vf2zz\") pod \"openstack-galera-0\" (UID: \"0195e48a-f8c9-433b-a9a8-c3da38457280\") " pod="openstack/openstack-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.952406 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.953764 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.955433 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.957149 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-hv8bp" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.957509 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.957787 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.961885 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 07:32:43 crc kubenswrapper[4829]: I1002 07:32:43.969423 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.084694 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.084744 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.084774 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.084805 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdf49\" (UniqueName: \"kubernetes.io/projected/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-kube-api-access-cdf49\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.084835 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.084993 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.085066 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.085213 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.085315 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.187162 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.187212 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.187265 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.187303 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdf49\" (UniqueName: \"kubernetes.io/projected/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-kube-api-access-cdf49\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188078 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188137 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188183 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188245 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188497 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188514 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.188551 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.189118 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.189419 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.191376 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.192951 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.207620 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.216014 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdf49\" (UniqueName: \"kubernetes.io/projected/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-kube-api-access-cdf49\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.218363 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.218766 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d72fc392-aa3b-40de-98cd-a0f7a633d5fa-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"d72fc392-aa3b-40de-98cd-a0f7a633d5fa\") " pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.300015 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.334347 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.335391 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.337123 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.338735 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-nzh8p" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.353427 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.353643 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.493421 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5146035b-0eb1-40c3-bc13-9d40affa0f2e-kolla-config\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.493545 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5c6k\" (UniqueName: \"kubernetes.io/projected/5146035b-0eb1-40c3-bc13-9d40affa0f2e-kube-api-access-c5c6k\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.493590 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5146035b-0eb1-40c3-bc13-9d40affa0f2e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.493630 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5146035b-0eb1-40c3-bc13-9d40affa0f2e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.493830 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5146035b-0eb1-40c3-bc13-9d40affa0f2e-config-data\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.595189 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5c6k\" (UniqueName: \"kubernetes.io/projected/5146035b-0eb1-40c3-bc13-9d40affa0f2e-kube-api-access-c5c6k\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.595259 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5146035b-0eb1-40c3-bc13-9d40affa0f2e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.595288 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5146035b-0eb1-40c3-bc13-9d40affa0f2e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.595369 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5146035b-0eb1-40c3-bc13-9d40affa0f2e-config-data\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.595417 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5146035b-0eb1-40c3-bc13-9d40affa0f2e-kolla-config\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.597825 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5146035b-0eb1-40c3-bc13-9d40affa0f2e-config-data\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.598035 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5146035b-0eb1-40c3-bc13-9d40affa0f2e-kolla-config\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.602430 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5146035b-0eb1-40c3-bc13-9d40affa0f2e-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.612717 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5146035b-0eb1-40c3-bc13-9d40affa0f2e-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.615271 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5c6k\" (UniqueName: \"kubernetes.io/projected/5146035b-0eb1-40c3-bc13-9d40affa0f2e-kube-api-access-c5c6k\") pod \"memcached-0\" (UID: \"5146035b-0eb1-40c3-bc13-9d40affa0f2e\") " pod="openstack/memcached-0" Oct 02 07:32:44 crc kubenswrapper[4829]: I1002 07:32:44.701617 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.159547 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.160517 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.163079 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-x6z5r" Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.174677 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.324152 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs9h5\" (UniqueName: \"kubernetes.io/projected/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3-kube-api-access-xs9h5\") pod \"kube-state-metrics-0\" (UID: \"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3\") " pod="openstack/kube-state-metrics-0" Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.425951 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs9h5\" (UniqueName: \"kubernetes.io/projected/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3-kube-api-access-xs9h5\") pod \"kube-state-metrics-0\" (UID: \"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3\") " pod="openstack/kube-state-metrics-0" Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.453152 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs9h5\" (UniqueName: \"kubernetes.io/projected/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3-kube-api-access-xs9h5\") pod \"kube-state-metrics-0\" (UID: \"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3\") " pod="openstack/kube-state-metrics-0" Oct 02 07:32:46 crc kubenswrapper[4829]: I1002 07:32:46.476618 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.482914 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.486888 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.488944 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.489272 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.489607 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.490372 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hrzwj" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.492256 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.504299 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.517741 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.647140 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.647687 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.647713 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.647985 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/61f93d4a-ce37-43ad-85d0-4be548d45c3b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.648014 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.648043 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2sgw\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-kube-api-access-k2sgw\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.648070 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.648093 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749436 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749546 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749668 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/61f93d4a-ce37-43ad-85d0-4be548d45c3b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749707 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749754 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2sgw\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-kube-api-access-k2sgw\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749807 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749843 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.749932 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.751559 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/61f93d4a-ce37-43ad-85d0-4be548d45c3b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.754588 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.755115 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.755721 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.756146 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.757962 4829 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.758005 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf460093261147c62473ff98f8811184b7db6218b040c193e53546be3326fb82/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.766706 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.781727 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2sgw\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-kube-api-access-k2sgw\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.803452 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:47 crc kubenswrapper[4829]: I1002 07:32:47.829287 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.201477 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-njw55"] Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.203028 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.205336 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-mlcpw" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.206084 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.206237 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.227257 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-njw55"] Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.243706 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-t8kr7"] Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.245727 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.254754 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t8kr7"] Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304010 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-log-ovn\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304051 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ff1e5f-73e2-4237-b201-84065a586553-combined-ca-bundle\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304093 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-run\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304219 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ff1e5f-73e2-4237-b201-84065a586553-ovn-controller-tls-certs\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304473 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-run-ovn\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304536 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91ff1e5f-73e2-4237-b201-84065a586553-scripts\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.304608 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh6sk\" (UniqueName: \"kubernetes.io/projected/91ff1e5f-73e2-4237-b201-84065a586553-kube-api-access-nh6sk\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.406671 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh6sk\" (UniqueName: \"kubernetes.io/projected/91ff1e5f-73e2-4237-b201-84065a586553-kube-api-access-nh6sk\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.406774 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-etc-ovs\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.406849 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-log-ovn\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.406912 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ff1e5f-73e2-4237-b201-84065a586553-combined-ca-bundle\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.406953 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-run\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407004 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ff1e5f-73e2-4237-b201-84065a586553-ovn-controller-tls-certs\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407036 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-run\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407080 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnrbh\" (UniqueName: \"kubernetes.io/projected/bcd90918-1fff-4cf6-9020-7780c2c53672-kube-api-access-bnrbh\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407178 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91ff1e5f-73e2-4237-b201-84065a586553-scripts\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407220 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-log\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407239 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-log-ovn\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407308 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-lib\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407449 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcd90918-1fff-4cf6-9020-7780c2c53672-scripts\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407556 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-run-ovn\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.407904 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-run\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.408048 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/91ff1e5f-73e2-4237-b201-84065a586553-var-run-ovn\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.410969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/91ff1e5f-73e2-4237-b201-84065a586553-scripts\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.411696 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/91ff1e5f-73e2-4237-b201-84065a586553-ovn-controller-tls-certs\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.420333 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91ff1e5f-73e2-4237-b201-84065a586553-combined-ca-bundle\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.424265 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh6sk\" (UniqueName: \"kubernetes.io/projected/91ff1e5f-73e2-4237-b201-84065a586553-kube-api-access-nh6sk\") pod \"ovn-controller-njw55\" (UID: \"91ff1e5f-73e2-4237-b201-84065a586553\") " pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509150 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-log\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509188 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-lib\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509235 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcd90918-1fff-4cf6-9020-7780c2c53672-scripts\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509289 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-etc-ovs\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509326 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-run\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509341 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnrbh\" (UniqueName: \"kubernetes.io/projected/bcd90918-1fff-4cf6-9020-7780c2c53672-kube-api-access-bnrbh\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.509953 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-lib\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.510032 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-run\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.510032 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-etc-ovs\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.511691 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bcd90918-1fff-4cf6-9020-7780c2c53672-scripts\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.511782 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/bcd90918-1fff-4cf6-9020-7780c2c53672-var-log\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.526444 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-njw55" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.527053 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnrbh\" (UniqueName: \"kubernetes.io/projected/bcd90918-1fff-4cf6-9020-7780c2c53672-kube-api-access-bnrbh\") pod \"ovn-controller-ovs-t8kr7\" (UID: \"bcd90918-1fff-4cf6-9020-7780c2c53672\") " pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:50 crc kubenswrapper[4829]: I1002 07:32:50.586873 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.205974 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.207580 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.209831 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.209982 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.210040 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.210755 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-t2m6h" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.210886 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.214997 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327000 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37f64531-608d-4dd5-99ad-5c928bcf7ca0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327467 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327630 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj477\" (UniqueName: \"kubernetes.io/projected/37f64531-608d-4dd5-99ad-5c928bcf7ca0-kube-api-access-kj477\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37f64531-608d-4dd5-99ad-5c928bcf7ca0-config\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327816 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327961 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37f64531-608d-4dd5-99ad-5c928bcf7ca0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.327991 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.328134 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429353 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37f64531-608d-4dd5-99ad-5c928bcf7ca0-config\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429419 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429443 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37f64531-608d-4dd5-99ad-5c928bcf7ca0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429464 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429486 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429532 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37f64531-608d-4dd5-99ad-5c928bcf7ca0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429584 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.429611 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj477\" (UniqueName: \"kubernetes.io/projected/37f64531-608d-4dd5-99ad-5c928bcf7ca0-kube-api-access-kj477\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.430570 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.430667 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37f64531-608d-4dd5-99ad-5c928bcf7ca0-config\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.431136 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/37f64531-608d-4dd5-99ad-5c928bcf7ca0-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.434124 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.435239 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.436779 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/37f64531-608d-4dd5-99ad-5c928bcf7ca0-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.447024 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/37f64531-608d-4dd5-99ad-5c928bcf7ca0-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.447472 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj477\" (UniqueName: \"kubernetes.io/projected/37f64531-608d-4dd5-99ad-5c928bcf7ca0-kube-api-access-kj477\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.459484 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"37f64531-608d-4dd5-99ad-5c928bcf7ca0\") " pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: I1002 07:32:51.526906 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 07:32:51 crc kubenswrapper[4829]: E1002 07:32:51.867474 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 07:32:51 crc kubenswrapper[4829]: E1002 07:32:51.867986 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qsvjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-ljpcp_openstack(a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:32:51 crc kubenswrapper[4829]: E1002 07:32:51.869317 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" podUID="a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3" Oct 02 07:32:51 crc kubenswrapper[4829]: E1002 07:32:51.877334 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 07:32:51 crc kubenswrapper[4829]: E1002 07:32:51.877495 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-trskc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-f2m9b_openstack(859cd9a2-b21a-4dd7-8649-97a677147c1d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:32:51 crc kubenswrapper[4829]: E1002 07:32:51.878693 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" podUID="859cd9a2-b21a-4dd7-8649-97a677147c1d" Oct 02 07:32:52 crc kubenswrapper[4829]: I1002 07:32:52.479579 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-mpllb"] Oct 02 07:32:52 crc kubenswrapper[4829]: I1002 07:32:52.505578 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fjdw2"] Oct 02 07:32:52 crc kubenswrapper[4829]: I1002 07:32:52.883103 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:52 crc kubenswrapper[4829]: I1002 07:32:52.897311 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:52 crc kubenswrapper[4829]: W1002 07:32:52.914694 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56bf3dfd_bae4_4af6_a338_fad9857bfd29.slice/crio-862ff59d1bbc09414b573d7a80d32493b48ffe9b180cc0aa329338c6e63ae7ee WatchSource:0}: Error finding container 862ff59d1bbc09414b573d7a80d32493b48ffe9b180cc0aa329338c6e63ae7ee: Status 404 returned error can't find the container with id 862ff59d1bbc09414b573d7a80d32493b48ffe9b180cc0aa329338c6e63ae7ee Oct 02 07:32:52 crc kubenswrapper[4829]: I1002 07:32:52.931217 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:32:52 crc kubenswrapper[4829]: I1002 07:32:52.948887 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.056582 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-dns-svc\") pod \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.056713 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trskc\" (UniqueName: \"kubernetes.io/projected/859cd9a2-b21a-4dd7-8649-97a677147c1d-kube-api-access-trskc\") pod \"859cd9a2-b21a-4dd7-8649-97a677147c1d\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.056755 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-config\") pod \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.056804 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859cd9a2-b21a-4dd7-8649-97a677147c1d-config\") pod \"859cd9a2-b21a-4dd7-8649-97a677147c1d\" (UID: \"859cd9a2-b21a-4dd7-8649-97a677147c1d\") " Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.056851 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvjs\" (UniqueName: \"kubernetes.io/projected/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-kube-api-access-qsvjs\") pod \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\" (UID: \"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3\") " Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.058112 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3" (UID: "a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.058442 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-config" (OuterVolumeSpecName: "config") pod "a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3" (UID: "a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.059018 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/859cd9a2-b21a-4dd7-8649-97a677147c1d-config" (OuterVolumeSpecName: "config") pod "859cd9a2-b21a-4dd7-8649-97a677147c1d" (UID: "859cd9a2-b21a-4dd7-8649-97a677147c1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.064238 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-kube-api-access-qsvjs" (OuterVolumeSpecName: "kube-api-access-qsvjs") pod "a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3" (UID: "a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3"). InnerVolumeSpecName "kube-api-access-qsvjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.065132 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/859cd9a2-b21a-4dd7-8649-97a677147c1d-kube-api-access-trskc" (OuterVolumeSpecName: "kube-api-access-trskc") pod "859cd9a2-b21a-4dd7-8649-97a677147c1d" (UID: "859cd9a2-b21a-4dd7-8649-97a677147c1d"). InnerVolumeSpecName "kube-api-access-trskc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.160849 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trskc\" (UniqueName: \"kubernetes.io/projected/859cd9a2-b21a-4dd7-8649-97a677147c1d-kube-api-access-trskc\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.160896 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.160915 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/859cd9a2-b21a-4dd7-8649-97a677147c1d-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.160927 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvjs\" (UniqueName: \"kubernetes.io/projected/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-kube-api-access-qsvjs\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.160938 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.165551 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.181489 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.194083 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.212694 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-njw55"] Oct 02 07:32:53 crc kubenswrapper[4829]: W1002 07:32:53.213613 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5146035b_0eb1_40c3_bc13_9d40affa0f2e.slice/crio-3794291727f5dfdc1f2331a51e3e052a47e1dd866012e6d2cc8e6b67225c2a2f WatchSource:0}: Error finding container 3794291727f5dfdc1f2331a51e3e052a47e1dd866012e6d2cc8e6b67225c2a2f: Status 404 returned error can't find the container with id 3794291727f5dfdc1f2331a51e3e052a47e1dd866012e6d2cc8e6b67225c2a2f Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.221180 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.230320 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.235764 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.237906 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.241751 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.242978 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.243162 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.243464 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-jl86t" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.248869 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.262753 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerStarted","Data":"ec5d5816bb3bb652b9384fb03f5b6bd8c18011da7d3a56b6e1e2380a796dbf0c"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.266107 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3","Type":"ContainerStarted","Data":"5c97ef80661509fa86c0121858c0373f14492b2c5af443228ac222800a22d06e"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.267395 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" event={"ID":"bab06794-e96a-49ea-b1c7-d99ae28c3630","Type":"ContainerStarted","Data":"dc97533e4ebcb0d0b034996c59b05cb30c008ea9aa197c9775a91cd7a7aed849"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.268557 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" event={"ID":"859cd9a2-b21a-4dd7-8649-97a677147c1d","Type":"ContainerDied","Data":"07164f20696e4eaa7624c41985e0351ed9ff2310873d8a02fb0aa0f00d6a239d"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.268619 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-f2m9b" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.270333 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-njw55" event={"ID":"91ff1e5f-73e2-4237-b201-84065a586553","Type":"ContainerStarted","Data":"75425195ba8435b509c63ad7ed6d0c31dc5ab4fec35f577e10fd6fbf03164a7b"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.272633 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0195e48a-f8c9-433b-a9a8-c3da38457280","Type":"ContainerStarted","Data":"ff1c2fe8c9ca52307573e8d3dab6eaa0a995b713b11f83f2354cd7b8d78927a7"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.273666 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" event={"ID":"9c93fa9a-4673-46f4-b432-ff794f518510","Type":"ContainerStarted","Data":"e77f5e2bb4c31856d08cfb03a917fa61a5b0cba6f62542c7f82915c07a7ba178"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.277071 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5146035b-0eb1-40c3-bc13-9d40affa0f2e","Type":"ContainerStarted","Data":"3794291727f5dfdc1f2331a51e3e052a47e1dd866012e6d2cc8e6b67225c2a2f"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.278369 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bf3dfd-bae4-4af6-a338-fad9857bfd29","Type":"ContainerStarted","Data":"862ff59d1bbc09414b573d7a80d32493b48ffe9b180cc0aa329338c6e63ae7ee"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.280790 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" event={"ID":"a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3","Type":"ContainerDied","Data":"044ac3cec8547a769a8ffac3e4eb2de4756f8b504cbc452467dc29ccf0a22c3d"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.280827 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-ljpcp" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.282612 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d72fc392-aa3b-40de-98cd-a0f7a633d5fa","Type":"ContainerStarted","Data":"f2a0312fd428f39bf687215d4c4de07590bd0daa42375593c4fdbe6886c04067"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.285646 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2","Type":"ContainerStarted","Data":"c1623bd0c6b05992aa4dfff70b38975e0735f2ca4df4b1a23138c7fb3220f8a7"} Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.324848 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.336140 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-f2m9b"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.343726 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-f2m9b"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.356333 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ljpcp"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.360661 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-ljpcp"] Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.363898 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.363961 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.364028 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.364125 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6gpp\" (UniqueName: \"kubernetes.io/projected/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-kube-api-access-c6gpp\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.365824 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.365856 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.365888 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-config\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.365928 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.466990 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6gpp\" (UniqueName: \"kubernetes.io/projected/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-kube-api-access-c6gpp\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467040 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467068 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467094 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-config\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467129 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467165 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467200 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.467225 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.468328 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.468841 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-config\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.469081 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.472753 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.473117 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.473492 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.482370 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.491018 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="859cd9a2-b21a-4dd7-8649-97a677147c1d" path="/var/lib/kubelet/pods/859cd9a2-b21a-4dd7-8649-97a677147c1d/volumes" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.491515 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3" path="/var/lib/kubelet/pods/a1ecf4ff-3aeb-4c5e-a4e3-095dc8c663b3/volumes" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.492503 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6gpp\" (UniqueName: \"kubernetes.io/projected/cedf4e9a-cfe0-4499-8600-edf90b9b5cda-kube-api-access-c6gpp\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.496042 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cedf4e9a-cfe0-4499-8600-edf90b9b5cda\") " pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:53 crc kubenswrapper[4829]: I1002 07:32:53.561196 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 07:32:54 crc kubenswrapper[4829]: I1002 07:32:54.057156 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 07:32:54 crc kubenswrapper[4829]: W1002 07:32:54.058314 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcedf4e9a_cfe0_4499_8600_edf90b9b5cda.slice/crio-016ba2748cfc4442c3a1746ce298bf2c16a50b9e0b3b0b4129cf476cb63c5b1d WatchSource:0}: Error finding container 016ba2748cfc4442c3a1746ce298bf2c16a50b9e0b3b0b4129cf476cb63c5b1d: Status 404 returned error can't find the container with id 016ba2748cfc4442c3a1746ce298bf2c16a50b9e0b3b0b4129cf476cb63c5b1d Oct 02 07:32:54 crc kubenswrapper[4829]: I1002 07:32:54.180385 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-t8kr7"] Oct 02 07:32:54 crc kubenswrapper[4829]: W1002 07:32:54.184432 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcd90918_1fff_4cf6_9020_7780c2c53672.slice/crio-5a30cf5672806cf5a1ea77feb5be981a9d7d6792632154cf231b50f2535a69f4 WatchSource:0}: Error finding container 5a30cf5672806cf5a1ea77feb5be981a9d7d6792632154cf231b50f2535a69f4: Status 404 returned error can't find the container with id 5a30cf5672806cf5a1ea77feb5be981a9d7d6792632154cf231b50f2535a69f4 Oct 02 07:32:54 crc kubenswrapper[4829]: I1002 07:32:54.303922 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cedf4e9a-cfe0-4499-8600-edf90b9b5cda","Type":"ContainerStarted","Data":"016ba2748cfc4442c3a1746ce298bf2c16a50b9e0b3b0b4129cf476cb63c5b1d"} Oct 02 07:32:54 crc kubenswrapper[4829]: I1002 07:32:54.310557 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t8kr7" event={"ID":"bcd90918-1fff-4cf6-9020-7780c2c53672","Type":"ContainerStarted","Data":"5a30cf5672806cf5a1ea77feb5be981a9d7d6792632154cf231b50f2535a69f4"} Oct 02 07:32:54 crc kubenswrapper[4829]: I1002 07:32:54.312263 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"37f64531-608d-4dd5-99ad-5c928bcf7ca0","Type":"ContainerStarted","Data":"e2be2fd5899028c7244b8c43c6fcef1c4e7802f81bd372461925a07ff42a2c6a"} Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.323118 4829 generic.go:334] "Generic (PLEG): container finished" podID="9c93fa9a-4673-46f4-b432-ff794f518510" containerID="2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918" exitCode=0 Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.323450 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" event={"ID":"9c93fa9a-4673-46f4-b432-ff794f518510","Type":"ContainerDied","Data":"2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918"} Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.327024 4829 generic.go:334] "Generic (PLEG): container finished" podID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerID="e753735c603b359da0f17f952a03aadc55003eca3d713443bb92670f7fde4cb3" exitCode=0 Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.327122 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" event={"ID":"bab06794-e96a-49ea-b1c7-d99ae28c3630","Type":"ContainerDied","Data":"e753735c603b359da0f17f952a03aadc55003eca3d713443bb92670f7fde4cb3"} Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.329466 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.329518 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.329557 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.330482 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"34b90026201c24a6d0bc1440495bb6980e11f5b356c27ed499750d010a61ec35"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:32:55 crc kubenswrapper[4829]: I1002 07:32:55.330555 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://34b90026201c24a6d0bc1440495bb6980e11f5b356c27ed499750d010a61ec35" gracePeriod=600 Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.335990 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" event={"ID":"9c93fa9a-4673-46f4-b432-ff794f518510","Type":"ContainerStarted","Data":"f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6"} Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.336595 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.342236 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="34b90026201c24a6d0bc1440495bb6980e11f5b356c27ed499750d010a61ec35" exitCode=0 Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.342267 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"34b90026201c24a6d0bc1440495bb6980e11f5b356c27ed499750d010a61ec35"} Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.342306 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"ae2b08236229e74b63b3e0438daf20903c2e9ff5c5e97936f06ade85ab34a7a5"} Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.342325 4829 scope.go:117] "RemoveContainer" containerID="d8e666d4540458f8abd1af2cd0bebf671777110d93b151c9aedce3fa191263d9" Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.346526 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" event={"ID":"bab06794-e96a-49ea-b1c7-d99ae28c3630","Type":"ContainerStarted","Data":"28afe37746ad135acbf3beaa5aaa5b9e4a9d4034e14ff087585a602d111494df"} Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.346679 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.353560 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" podStartSLOduration=15.197678473 podStartE2EDuration="17.353544868s" podCreationTimestamp="2025-10-02 07:32:39 +0000 UTC" firstStartedPulling="2025-10-02 07:32:52.511883399 +0000 UTC m=+963.851531794" lastFinishedPulling="2025-10-02 07:32:54.667749774 +0000 UTC m=+966.007398189" observedRunningTime="2025-10-02 07:32:56.351015867 +0000 UTC m=+967.690664272" watchObservedRunningTime="2025-10-02 07:32:56.353544868 +0000 UTC m=+967.693193273" Oct 02 07:32:56 crc kubenswrapper[4829]: I1002 07:32:56.380815 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" podStartSLOduration=15.102680023 podStartE2EDuration="17.380794479s" podCreationTimestamp="2025-10-02 07:32:39 +0000 UTC" firstStartedPulling="2025-10-02 07:32:52.477194568 +0000 UTC m=+963.816842973" lastFinishedPulling="2025-10-02 07:32:54.755309014 +0000 UTC m=+966.094957429" observedRunningTime="2025-10-02 07:32:56.377711309 +0000 UTC m=+967.717359714" watchObservedRunningTime="2025-10-02 07:32:56.380794479 +0000 UTC m=+967.720442884" Oct 02 07:33:00 crc kubenswrapper[4829]: I1002 07:33:00.224121 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:33:00 crc kubenswrapper[4829]: I1002 07:33:00.271894 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-mpllb"] Oct 02 07:33:00 crc kubenswrapper[4829]: I1002 07:33:00.273384 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerName="dnsmasq-dns" containerID="cri-o://28afe37746ad135acbf3beaa5aaa5b9e4a9d4034e14ff087585a602d111494df" gracePeriod=10 Oct 02 07:33:00 crc kubenswrapper[4829]: I1002 07:33:00.277416 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:33:01 crc kubenswrapper[4829]: I1002 07:33:01.379064 4829 generic.go:334] "Generic (PLEG): container finished" podID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerID="28afe37746ad135acbf3beaa5aaa5b9e4a9d4034e14ff087585a602d111494df" exitCode=0 Oct 02 07:33:01 crc kubenswrapper[4829]: I1002 07:33:01.379104 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" event={"ID":"bab06794-e96a-49ea-b1c7-d99ae28c3630","Type":"ContainerDied","Data":"28afe37746ad135acbf3beaa5aaa5b9e4a9d4034e14ff087585a602d111494df"} Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.328356 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.394566 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.394563 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-mpllb" event={"ID":"bab06794-e96a-49ea-b1c7-d99ae28c3630","Type":"ContainerDied","Data":"dc97533e4ebcb0d0b034996c59b05cb30c008ea9aa197c9775a91cd7a7aed849"} Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.395134 4829 scope.go:117] "RemoveContainer" containerID="28afe37746ad135acbf3beaa5aaa5b9e4a9d4034e14ff087585a602d111494df" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.455432 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-config\") pod \"bab06794-e96a-49ea-b1c7-d99ae28c3630\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.455477 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzgc6\" (UniqueName: \"kubernetes.io/projected/bab06794-e96a-49ea-b1c7-d99ae28c3630-kube-api-access-mzgc6\") pod \"bab06794-e96a-49ea-b1c7-d99ae28c3630\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.455550 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-dns-svc\") pod \"bab06794-e96a-49ea-b1c7-d99ae28c3630\" (UID: \"bab06794-e96a-49ea-b1c7-d99ae28c3630\") " Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.458442 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bab06794-e96a-49ea-b1c7-d99ae28c3630-kube-api-access-mzgc6" (OuterVolumeSpecName: "kube-api-access-mzgc6") pod "bab06794-e96a-49ea-b1c7-d99ae28c3630" (UID: "bab06794-e96a-49ea-b1c7-d99ae28c3630"). InnerVolumeSpecName "kube-api-access-mzgc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.507478 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bab06794-e96a-49ea-b1c7-d99ae28c3630" (UID: "bab06794-e96a-49ea-b1c7-d99ae28c3630"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.509590 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-config" (OuterVolumeSpecName: "config") pod "bab06794-e96a-49ea-b1c7-d99ae28c3630" (UID: "bab06794-e96a-49ea-b1c7-d99ae28c3630"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.524115 4829 scope.go:117] "RemoveContainer" containerID="e753735c603b359da0f17f952a03aadc55003eca3d713443bb92670f7fde4cb3" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.557808 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.557849 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bab06794-e96a-49ea-b1c7-d99ae28c3630-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.557861 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzgc6\" (UniqueName: \"kubernetes.io/projected/bab06794-e96a-49ea-b1c7-d99ae28c3630-kube-api-access-mzgc6\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.725820 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-mpllb"] Oct 02 07:33:03 crc kubenswrapper[4829]: I1002 07:33:03.731579 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-mpllb"] Oct 02 07:33:04 crc kubenswrapper[4829]: I1002 07:33:04.410537 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cedf4e9a-cfe0-4499-8600-edf90b9b5cda","Type":"ContainerStarted","Data":"350717d9e16cf9c0346e3b65d75c3b8da20bc53ea6f32b18b0692c2902e37593"} Oct 02 07:33:05 crc kubenswrapper[4829]: I1002 07:33:05.469661 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" path="/var/lib/kubelet/pods/bab06794-e96a-49ea-b1c7-d99ae28c3630/volumes" Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.440221 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d72fc392-aa3b-40de-98cd-a0f7a633d5fa","Type":"ContainerStarted","Data":"bee6a7540688aaed9e4cfd54e6797991e8fa28eda970b00ea3fe29b7aa2a51f7"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.443278 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2","Type":"ContainerStarted","Data":"3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.449921 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5146035b-0eb1-40c3-bc13-9d40affa0f2e","Type":"ContainerStarted","Data":"b3e5db55746fc802242246fd55e01f9f6b7ae2d51c55c4a04af796208b5c8e60"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.450343 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.452304 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-njw55" event={"ID":"91ff1e5f-73e2-4237-b201-84065a586553","Type":"ContainerStarted","Data":"dc01ea1e5cd80ec8304e12778d2b87df2c13cdfae49f97fdd21d10ce84a32ba3"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.452910 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-njw55" Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.458441 4829 generic.go:334] "Generic (PLEG): container finished" podID="bcd90918-1fff-4cf6-9020-7780c2c53672" containerID="0c83a5bbb7a92c1869c61f1381229bfe3924100473ffcf43aa5f4ca39ea2ddbe" exitCode=0 Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.458517 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t8kr7" event={"ID":"bcd90918-1fff-4cf6-9020-7780c2c53672","Type":"ContainerDied","Data":"0c83a5bbb7a92c1869c61f1381229bfe3924100473ffcf43aa5f4ca39ea2ddbe"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.461861 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"37f64531-608d-4dd5-99ad-5c928bcf7ca0","Type":"ContainerStarted","Data":"e8003c7d6c7526f619596398d35ff413e97ad613bc9298c6e9574a3d390b1407"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.464459 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3","Type":"ContainerStarted","Data":"76959c27443e03cc3089c6da40ef4e114f4c5d9bd80af46e12cb7814d888978f"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.465142 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.471307 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bf3dfd-bae4-4af6-a338-fad9857bfd29","Type":"ContainerStarted","Data":"fad8f75f06da64a1030f412dfcd4668bdcec0ea2d6b25739b69ddbef0345d357"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.474590 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0195e48a-f8c9-433b-a9a8-c3da38457280","Type":"ContainerStarted","Data":"5ba5972d7b6963f5c05323ccb036f62456e731ce280f30d0da221f992db53387"} Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.504777 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.515399065 podStartE2EDuration="20.504758143s" podCreationTimestamp="2025-10-02 07:32:46 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.194324492 +0000 UTC m=+964.533972907" lastFinishedPulling="2025-10-02 07:33:04.18368356 +0000 UTC m=+975.523331985" observedRunningTime="2025-10-02 07:33:06.498292014 +0000 UTC m=+977.837940439" watchObservedRunningTime="2025-10-02 07:33:06.504758143 +0000 UTC m=+977.844406548" Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.524463 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-njw55" podStartSLOduration=6.284222939 podStartE2EDuration="16.52444183s" podCreationTimestamp="2025-10-02 07:32:50 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.195602413 +0000 UTC m=+964.535250808" lastFinishedPulling="2025-10-02 07:33:03.435821294 +0000 UTC m=+974.775469699" observedRunningTime="2025-10-02 07:33:06.51579195 +0000 UTC m=+977.855440365" watchObservedRunningTime="2025-10-02 07:33:06.52444183 +0000 UTC m=+977.864090235" Oct 02 07:33:06 crc kubenswrapper[4829]: I1002 07:33:06.569221 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.934911125 podStartE2EDuration="22.569206466s" podCreationTimestamp="2025-10-02 07:32:44 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.217315185 +0000 UTC m=+964.556963590" lastFinishedPulling="2025-10-02 07:33:02.851610506 +0000 UTC m=+974.191258931" observedRunningTime="2025-10-02 07:33:06.566813099 +0000 UTC m=+977.906461524" watchObservedRunningTime="2025-10-02 07:33:06.569206466 +0000 UTC m=+977.908854871" Oct 02 07:33:07 crc kubenswrapper[4829]: I1002 07:33:07.495816 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerStarted","Data":"a8d3d630c1db387878ae8bb1b1baed1a5515ac74b05ca5b0f1313dc3793d0171"} Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.511725 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cedf4e9a-cfe0-4499-8600-edf90b9b5cda","Type":"ContainerStarted","Data":"48fd8bcfdd5715b008236db38c74b6961685693828bc5485e87338f772a215c9"} Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.515168 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t8kr7" event={"ID":"bcd90918-1fff-4cf6-9020-7780c2c53672","Type":"ContainerStarted","Data":"256b9e94d0021dc42f3f70a5342c38908c96cebaf6323b71e785bb48ce7da29d"} Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.515199 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-t8kr7" event={"ID":"bcd90918-1fff-4cf6-9020-7780c2c53672","Type":"ContainerStarted","Data":"6278ffa6e87d0acdcd585e6e87fe9296a1e119fa45441673c962d037f93d4348"} Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.515730 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.515767 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.518138 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"37f64531-608d-4dd5-99ad-5c928bcf7ca0","Type":"ContainerStarted","Data":"8a7cbfb7ee72ae1d3468a8d7a0c3a1c9c7f2a85dcfeacdab5901d122e79ed84c"} Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.561967 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.562027 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.577379 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.695072586 podStartE2EDuration="16.577353366s" podCreationTimestamp="2025-10-02 07:32:52 +0000 UTC" firstStartedPulling="2025-10-02 07:32:54.060406668 +0000 UTC m=+965.400055093" lastFinishedPulling="2025-10-02 07:33:07.942687458 +0000 UTC m=+979.282335873" observedRunningTime="2025-10-02 07:33:08.570187935 +0000 UTC m=+979.909836370" watchObservedRunningTime="2025-10-02 07:33:08.577353366 +0000 UTC m=+979.917001781" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.596437 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.97969724 podStartE2EDuration="18.596411963s" podCreationTimestamp="2025-10-02 07:32:50 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.324432116 +0000 UTC m=+964.664080521" lastFinishedPulling="2025-10-02 07:33:07.941146849 +0000 UTC m=+979.280795244" observedRunningTime="2025-10-02 07:33:08.58858473 +0000 UTC m=+979.928233155" watchObservedRunningTime="2025-10-02 07:33:08.596411963 +0000 UTC m=+979.936060378" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.611719 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-t8kr7" podStartSLOduration=9.284970534 podStartE2EDuration="18.611698207s" podCreationTimestamp="2025-10-02 07:32:50 +0000 UTC" firstStartedPulling="2025-10-02 07:32:54.186483052 +0000 UTC m=+965.526131457" lastFinishedPulling="2025-10-02 07:33:03.513210705 +0000 UTC m=+974.852859130" observedRunningTime="2025-10-02 07:33:08.607287395 +0000 UTC m=+979.946935850" watchObservedRunningTime="2025-10-02 07:33:08.611698207 +0000 UTC m=+979.951346622" Oct 02 07:33:08 crc kubenswrapper[4829]: I1002 07:33:08.626329 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.527717 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.530314 4829 generic.go:334] "Generic (PLEG): container finished" podID="0195e48a-f8c9-433b-a9a8-c3da38457280" containerID="5ba5972d7b6963f5c05323ccb036f62456e731ce280f30d0da221f992db53387" exitCode=0 Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.530394 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0195e48a-f8c9-433b-a9a8-c3da38457280","Type":"ContainerDied","Data":"5ba5972d7b6963f5c05323ccb036f62456e731ce280f30d0da221f992db53387"} Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.533855 4829 generic.go:334] "Generic (PLEG): container finished" podID="d72fc392-aa3b-40de-98cd-a0f7a633d5fa" containerID="bee6a7540688aaed9e4cfd54e6797991e8fa28eda970b00ea3fe29b7aa2a51f7" exitCode=0 Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.534182 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d72fc392-aa3b-40de-98cd-a0f7a633d5fa","Type":"ContainerDied","Data":"bee6a7540688aaed9e4cfd54e6797991e8fa28eda970b00ea3fe29b7aa2a51f7"} Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.614049 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.626386 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.957104 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p2p2c"] Oct 02 07:33:09 crc kubenswrapper[4829]: E1002 07:33:09.958133 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerName="dnsmasq-dns" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.958152 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerName="dnsmasq-dns" Oct 02 07:33:09 crc kubenswrapper[4829]: E1002 07:33:09.958170 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerName="init" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.958177 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerName="init" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.958519 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab06794-e96a-49ea-b1c7-d99ae28c3630" containerName="dnsmasq-dns" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.963148 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.967894 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p2p2c"] Oct 02 07:33:09 crc kubenswrapper[4829]: I1002 07:33:09.968381 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.021719 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-s5fpj"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.034923 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.039046 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.041459 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-s5fpj"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095313 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095353 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgst6\" (UniqueName: \"kubernetes.io/projected/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-kube-api-access-qgst6\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095373 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-ovs-rundir\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095393 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095411 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf4bv\" (UniqueName: \"kubernetes.io/projected/f989ac9a-e5b3-45c1-9999-7702b489e664-kube-api-access-hf4bv\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095627 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095685 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-combined-ca-bundle\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095742 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-config\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095772 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-config\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.095811 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-ovn-rundir\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.163780 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p2p2c"] Oct 02 07:33:10 crc kubenswrapper[4829]: E1002 07:33:10.164443 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-hf4bv ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" podUID="f989ac9a-e5b3-45c1-9999-7702b489e664" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.192979 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dmrvn"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.194462 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.196069 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.196865 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.196895 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgst6\" (UniqueName: \"kubernetes.io/projected/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-kube-api-access-qgst6\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.196918 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-ovs-rundir\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.196936 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.196953 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf4bv\" (UniqueName: \"kubernetes.io/projected/f989ac9a-e5b3-45c1-9999-7702b489e664-kube-api-access-hf4bv\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197008 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197024 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-combined-ca-bundle\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197047 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-config\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197063 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-config\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197084 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-ovn-rundir\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197344 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-ovn-rundir\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.197486 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-ovs-rundir\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.198514 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.203764 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-config\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.204238 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.204545 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-combined-ca-bundle\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.204882 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.208796 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-config\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.211170 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dmrvn"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.216927 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf4bv\" (UniqueName: \"kubernetes.io/projected/f989ac9a-e5b3-45c1-9999-7702b489e664-kube-api-access-hf4bv\") pod \"dnsmasq-dns-7f896c8c65-p2p2c\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.243470 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgst6\" (UniqueName: \"kubernetes.io/projected/4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3-kube-api-access-qgst6\") pod \"ovn-controller-metrics-s5fpj\" (UID: \"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3\") " pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.298371 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.298411 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x697\" (UniqueName: \"kubernetes.io/projected/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-kube-api-access-4x697\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.298442 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.298469 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-config\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.298485 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.365487 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-s5fpj" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.400551 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-config\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.400589 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.400700 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.400719 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x697\" (UniqueName: \"kubernetes.io/projected/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-kube-api-access-4x697\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.400744 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.401475 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.401550 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-config\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.401871 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.402493 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.416850 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x697\" (UniqueName: \"kubernetes.io/projected/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-kube-api-access-4x697\") pod \"dnsmasq-dns-86db49b7ff-dmrvn\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.545588 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0195e48a-f8c9-433b-a9a8-c3da38457280","Type":"ContainerStarted","Data":"d2063398c06a757e5c0ab416f4054ae3ad56ea339168bc527fd14f9d4724a8ee"} Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.551498 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d72fc392-aa3b-40de-98cd-a0f7a633d5fa","Type":"ContainerStarted","Data":"0ea73357935561b36d3782966d6566194f6f078f2f0c423c98aa84b8ab2cb206"} Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.551631 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.552134 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.578621 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.159363298 podStartE2EDuration="28.578602265s" podCreationTimestamp="2025-10-02 07:32:42 +0000 UTC" firstStartedPulling="2025-10-02 07:32:52.930595419 +0000 UTC m=+964.270243834" lastFinishedPulling="2025-10-02 07:33:03.349834386 +0000 UTC m=+974.689482801" observedRunningTime="2025-10-02 07:33:10.567335952 +0000 UTC m=+981.906984367" watchObservedRunningTime="2025-10-02 07:33:10.578602265 +0000 UTC m=+981.918250670" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.584405 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.595751 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.156174536 podStartE2EDuration="28.595737449s" podCreationTimestamp="2025-10-02 07:32:42 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.147213109 +0000 UTC m=+964.486861534" lastFinishedPulling="2025-10-02 07:33:03.586776032 +0000 UTC m=+974.926424447" observedRunningTime="2025-10-02 07:33:10.595162801 +0000 UTC m=+981.934811216" watchObservedRunningTime="2025-10-02 07:33:10.595737449 +0000 UTC m=+981.935385854" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.613189 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.621789 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.706115 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-dns-svc\") pod \"f989ac9a-e5b3-45c1-9999-7702b489e664\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.706468 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-config\") pod \"f989ac9a-e5b3-45c1-9999-7702b489e664\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.706555 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf4bv\" (UniqueName: \"kubernetes.io/projected/f989ac9a-e5b3-45c1-9999-7702b489e664-kube-api-access-hf4bv\") pod \"f989ac9a-e5b3-45c1-9999-7702b489e664\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.706605 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-ovsdbserver-sb\") pod \"f989ac9a-e5b3-45c1-9999-7702b489e664\" (UID: \"f989ac9a-e5b3-45c1-9999-7702b489e664\") " Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.707341 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f989ac9a-e5b3-45c1-9999-7702b489e664" (UID: "f989ac9a-e5b3-45c1-9999-7702b489e664"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.707593 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f989ac9a-e5b3-45c1-9999-7702b489e664" (UID: "f989ac9a-e5b3-45c1-9999-7702b489e664"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.707804 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-config" (OuterVolumeSpecName: "config") pod "f989ac9a-e5b3-45c1-9999-7702b489e664" (UID: "f989ac9a-e5b3-45c1-9999-7702b489e664"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.711431 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f989ac9a-e5b3-45c1-9999-7702b489e664-kube-api-access-hf4bv" (OuterVolumeSpecName: "kube-api-access-hf4bv") pod "f989ac9a-e5b3-45c1-9999-7702b489e664" (UID: "f989ac9a-e5b3-45c1-9999-7702b489e664"). InnerVolumeSpecName "kube-api-access-hf4bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.808519 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.808548 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf4bv\" (UniqueName: \"kubernetes.io/projected/f989ac9a-e5b3-45c1-9999-7702b489e664-kube-api-access-hf4bv\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.808562 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.808572 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f989ac9a-e5b3-45c1-9999-7702b489e664-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.862860 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.864387 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.870878 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.871036 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-t4j6m" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.871098 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.871430 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.873920 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-s5fpj"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.886325 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910655 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910782 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775c00d4-e88e-42a8-9351-edd61e0c2cd6-config\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910832 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/775c00d4-e88e-42a8-9351-edd61e0c2cd6-scripts\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910883 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsw6c\" (UniqueName: \"kubernetes.io/projected/775c00d4-e88e-42a8-9351-edd61e0c2cd6-kube-api-access-lsw6c\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910904 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910928 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/775c00d4-e88e-42a8-9351-edd61e0c2cd6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:10 crc kubenswrapper[4829]: I1002 07:33:10.910955 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.012810 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.012923 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775c00d4-e88e-42a8-9351-edd61e0c2cd6-config\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.013035 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/775c00d4-e88e-42a8-9351-edd61e0c2cd6-scripts\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.013986 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/775c00d4-e88e-42a8-9351-edd61e0c2cd6-scripts\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.014003 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/775c00d4-e88e-42a8-9351-edd61e0c2cd6-config\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.014102 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsw6c\" (UniqueName: \"kubernetes.io/projected/775c00d4-e88e-42a8-9351-edd61e0c2cd6-kube-api-access-lsw6c\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.014162 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.014689 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/775c00d4-e88e-42a8-9351-edd61e0c2cd6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.014728 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.015254 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/775c00d4-e88e-42a8-9351-edd61e0c2cd6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.019153 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.020201 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.021802 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/775c00d4-e88e-42a8-9351-edd61e0c2cd6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.029047 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsw6c\" (UniqueName: \"kubernetes.io/projected/775c00d4-e88e-42a8-9351-edd61e0c2cd6-kube-api-access-lsw6c\") pod \"ovn-northd-0\" (UID: \"775c00d4-e88e-42a8-9351-edd61e0c2cd6\") " pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.141056 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dmrvn"] Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.190040 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.561942 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-s5fpj" event={"ID":"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3","Type":"ContainerStarted","Data":"95f9ae9a5d328520fdf787f081ec9a432fb9ab39ac132c8796354ff13f4b5c7c"} Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.562271 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-s5fpj" event={"ID":"4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3","Type":"ContainerStarted","Data":"fca647d9477fb0bbe009e64bfb1ff0016a7508578c0397917d6a39e150e521de"} Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.573465 4829 generic.go:334] "Generic (PLEG): container finished" podID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerID="33c62eb6e94f2adc8f0e0863e6aab1d0c33debbca0d6bc169b4a8c7b96eabc63" exitCode=0 Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.574248 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" event={"ID":"c1aa7752-d0d1-4dd9-9a49-8aee61349f10","Type":"ContainerDied","Data":"33c62eb6e94f2adc8f0e0863e6aab1d0c33debbca0d6bc169b4a8c7b96eabc63"} Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.574280 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" event={"ID":"c1aa7752-d0d1-4dd9-9a49-8aee61349f10","Type":"ContainerStarted","Data":"fdc0c275d44680f34c5896db231dc6fb365119fe362f7298cadde1125217c7fb"} Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.574523 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-p2p2c" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.587629 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-s5fpj" podStartSLOduration=2.5876124799999998 podStartE2EDuration="2.58761248s" podCreationTimestamp="2025-10-02 07:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:11.581839073 +0000 UTC m=+982.921487488" watchObservedRunningTime="2025-10-02 07:33:11.58761248 +0000 UTC m=+982.927260885" Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.682150 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p2p2c"] Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.693423 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-p2p2c"] Oct 02 07:33:11 crc kubenswrapper[4829]: I1002 07:33:11.698932 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 07:33:12 crc kubenswrapper[4829]: I1002 07:33:12.590574 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"775c00d4-e88e-42a8-9351-edd61e0c2cd6","Type":"ContainerStarted","Data":"e948b0e702290376f34efb566561c2eeb98e98a98d07e0d2a0d2db5ad70c9787"} Oct 02 07:33:12 crc kubenswrapper[4829]: I1002 07:33:12.595900 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" event={"ID":"c1aa7752-d0d1-4dd9-9a49-8aee61349f10","Type":"ContainerStarted","Data":"c7756e81f5c0bc6745c8a8fcf9502fd3e1057a68615797c8417cb9d226fc49eb"} Oct 02 07:33:12 crc kubenswrapper[4829]: I1002 07:33:12.596486 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:12 crc kubenswrapper[4829]: I1002 07:33:12.625023 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" podStartSLOduration=2.625004282 podStartE2EDuration="2.625004282s" podCreationTimestamp="2025-10-02 07:33:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:12.621589872 +0000 UTC m=+983.961238287" watchObservedRunningTime="2025-10-02 07:33:12.625004282 +0000 UTC m=+983.964652687" Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.479653 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f989ac9a-e5b3-45c1-9999-7702b489e664" path="/var/lib/kubelet/pods/f989ac9a-e5b3-45c1-9999-7702b489e664/volumes" Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.608724 4829 generic.go:334] "Generic (PLEG): container finished" podID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerID="a8d3d630c1db387878ae8bb1b1baed1a5515ac74b05ca5b0f1313dc3793d0171" exitCode=0 Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.608811 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerDied","Data":"a8d3d630c1db387878ae8bb1b1baed1a5515ac74b05ca5b0f1313dc3793d0171"} Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.615897 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"775c00d4-e88e-42a8-9351-edd61e0c2cd6","Type":"ContainerStarted","Data":"783b82bbf02b484350e30000f50fc2d031724344d1c4730da2f1d12968658bde"} Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.616191 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"775c00d4-e88e-42a8-9351-edd61e0c2cd6","Type":"ContainerStarted","Data":"e3786f7998a12ddd8d0bfcd0fa5f622ae68ffa307635c6c81a0e29bd7a171c5d"} Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.616208 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.669257 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.514512442 podStartE2EDuration="3.669213605s" podCreationTimestamp="2025-10-02 07:33:10 +0000 UTC" firstStartedPulling="2025-10-02 07:33:11.703256787 +0000 UTC m=+983.042905202" lastFinishedPulling="2025-10-02 07:33:12.85795796 +0000 UTC m=+984.197606365" observedRunningTime="2025-10-02 07:33:13.665959589 +0000 UTC m=+985.005608014" watchObservedRunningTime="2025-10-02 07:33:13.669213605 +0000 UTC m=+985.008862020" Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.970962 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 07:33:13 crc kubenswrapper[4829]: I1002 07:33:13.971012 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 07:33:14 crc kubenswrapper[4829]: I1002 07:33:14.301564 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 07:33:14 crc kubenswrapper[4829]: I1002 07:33:14.301896 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 07:33:14 crc kubenswrapper[4829]: I1002 07:33:14.703715 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.498661 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.555497 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dmrvn"] Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.555699 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="dnsmasq-dns" containerID="cri-o://c7756e81f5c0bc6745c8a8fcf9502fd3e1057a68615797c8417cb9d226fc49eb" gracePeriod=10 Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.582411 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-kj4jn"] Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.582729 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: connect: connection refused" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.588679 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.597008 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kj4jn"] Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.711153 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.711267 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-config\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.711295 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5dsl\" (UniqueName: \"kubernetes.io/projected/5919e153-3626-4764-8110-ae849b21ff17-kube-api-access-w5dsl\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.711403 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-dns-svc\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.711449 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.812592 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.812692 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.812745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-config\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.812767 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5dsl\" (UniqueName: \"kubernetes.io/projected/5919e153-3626-4764-8110-ae849b21ff17-kube-api-access-w5dsl\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.812813 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-dns-svc\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.813656 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.813749 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-dns-svc\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.814105 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.814343 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-config\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.834145 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5dsl\" (UniqueName: \"kubernetes.io/projected/5919e153-3626-4764-8110-ae849b21ff17-kube-api-access-w5dsl\") pod \"dnsmasq-dns-698758b865-kj4jn\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:16 crc kubenswrapper[4829]: I1002 07:33:16.908850 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.661313 4829 generic.go:334] "Generic (PLEG): container finished" podID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerID="c7756e81f5c0bc6745c8a8fcf9502fd3e1057a68615797c8417cb9d226fc49eb" exitCode=0 Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.661398 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" event={"ID":"c1aa7752-d0d1-4dd9-9a49-8aee61349f10","Type":"ContainerDied","Data":"c7756e81f5c0bc6745c8a8fcf9502fd3e1057a68615797c8417cb9d226fc49eb"} Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.720085 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.736332 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.746995 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-28r2j" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.747727 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.748589 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.751944 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.765059 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.829484 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.829564 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txdjs\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-kube-api-access-txdjs\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.829622 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-lock\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.829793 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-cache\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.829899 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.931367 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txdjs\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-kube-api-access-txdjs\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.931422 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-lock\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.931467 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-cache\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.931494 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.931544 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: E1002 07:33:17.931669 4829 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 07:33:17 crc kubenswrapper[4829]: E1002 07:33:17.931683 4829 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 07:33:17 crc kubenswrapper[4829]: E1002 07:33:17.931717 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift podName:9da44bca-624e-49ca-8fd2-e0b0974c9ae5 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:18.431704032 +0000 UTC m=+989.771352437 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift") pod "swift-storage-0" (UID: "9da44bca-624e-49ca-8fd2-e0b0974c9ae5") : configmap "swift-ring-files" not found Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.932057 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.932163 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-lock\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.932522 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-cache\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.955659 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:17 crc kubenswrapper[4829]: I1002 07:33:17.958185 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txdjs\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-kube-api-access-txdjs\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:18 crc kubenswrapper[4829]: I1002 07:33:18.085020 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 07:33:18 crc kubenswrapper[4829]: I1002 07:33:18.144930 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 07:33:18 crc kubenswrapper[4829]: I1002 07:33:18.438819 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:18 crc kubenswrapper[4829]: E1002 07:33:18.439020 4829 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 07:33:18 crc kubenswrapper[4829]: E1002 07:33:18.439043 4829 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 07:33:18 crc kubenswrapper[4829]: E1002 07:33:18.439112 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift podName:9da44bca-624e-49ca-8fd2-e0b0974c9ae5 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:19.439094798 +0000 UTC m=+990.778743203 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift") pod "swift-storage-0" (UID: "9da44bca-624e-49ca-8fd2-e0b0974c9ae5") : configmap "swift-ring-files" not found Oct 02 07:33:18 crc kubenswrapper[4829]: I1002 07:33:18.887512 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 07:33:18 crc kubenswrapper[4829]: I1002 07:33:18.964430 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 07:33:19 crc kubenswrapper[4829]: I1002 07:33:19.134030 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kj4jn"] Oct 02 07:33:19 crc kubenswrapper[4829]: I1002 07:33:19.458191 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:19 crc kubenswrapper[4829]: E1002 07:33:19.458377 4829 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 07:33:19 crc kubenswrapper[4829]: E1002 07:33:19.458550 4829 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 07:33:19 crc kubenswrapper[4829]: E1002 07:33:19.458612 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift podName:9da44bca-624e-49ca-8fd2-e0b0974c9ae5 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:21.458598422 +0000 UTC m=+992.798246827 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift") pod "swift-storage-0" (UID: "9da44bca-624e-49ca-8fd2-e0b0974c9ae5") : configmap "swift-ring-files" not found Oct 02 07:33:19 crc kubenswrapper[4829]: I1002 07:33:19.684704 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kj4jn" event={"ID":"5919e153-3626-4764-8110-ae849b21ff17","Type":"ContainerStarted","Data":"7b5071013f9d8a59e3b0b9433a27e679ce6ae456aaf1cea747161df2a21e0cea"} Oct 02 07:33:20 crc kubenswrapper[4829]: I1002 07:33:20.623768 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.120:5353: connect: connection refused" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.172431 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.297168 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-dns-svc\") pod \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.297289 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-nb\") pod \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.297328 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x697\" (UniqueName: \"kubernetes.io/projected/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-kube-api-access-4x697\") pod \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.297371 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-config\") pod \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.297398 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-sb\") pod \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\" (UID: \"c1aa7752-d0d1-4dd9-9a49-8aee61349f10\") " Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.311468 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-kube-api-access-4x697" (OuterVolumeSpecName: "kube-api-access-4x697") pod "c1aa7752-d0d1-4dd9-9a49-8aee61349f10" (UID: "c1aa7752-d0d1-4dd9-9a49-8aee61349f10"). InnerVolumeSpecName "kube-api-access-4x697". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.364323 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-config" (OuterVolumeSpecName: "config") pod "c1aa7752-d0d1-4dd9-9a49-8aee61349f10" (UID: "c1aa7752-d0d1-4dd9-9a49-8aee61349f10"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.366212 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c1aa7752-d0d1-4dd9-9a49-8aee61349f10" (UID: "c1aa7752-d0d1-4dd9-9a49-8aee61349f10"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.368915 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c1aa7752-d0d1-4dd9-9a49-8aee61349f10" (UID: "c1aa7752-d0d1-4dd9-9a49-8aee61349f10"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.370469 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c1aa7752-d0d1-4dd9-9a49-8aee61349f10" (UID: "c1aa7752-d0d1-4dd9-9a49-8aee61349f10"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.401553 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.401598 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.401616 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x697\" (UniqueName: \"kubernetes.io/projected/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-kube-api-access-4x697\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.401629 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.401640 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c1aa7752-d0d1-4dd9-9a49-8aee61349f10-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.517592 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:21 crc kubenswrapper[4829]: E1002 07:33:21.517913 4829 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 07:33:21 crc kubenswrapper[4829]: E1002 07:33:21.517936 4829 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 07:33:21 crc kubenswrapper[4829]: E1002 07:33:21.517998 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift podName:9da44bca-624e-49ca-8fd2-e0b0974c9ae5 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:25.517974499 +0000 UTC m=+996.857622914 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift") pod "swift-storage-0" (UID: "9da44bca-624e-49ca-8fd2-e0b0974c9ae5") : configmap "swift-ring-files" not found Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.663283 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-2t6js"] Oct 02 07:33:21 crc kubenswrapper[4829]: E1002 07:33:21.663973 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="init" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.663991 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="init" Oct 02 07:33:21 crc kubenswrapper[4829]: E1002 07:33:21.664026 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="dnsmasq-dns" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.664034 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="dnsmasq-dns" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.673209 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" containerName="dnsmasq-dns" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.674037 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.675369 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-2t6js"] Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.675974 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.676010 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.676096 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.708533 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2t6js"] Oct 02 07:33:21 crc kubenswrapper[4829]: E1002 07:33:21.715202 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-5kmr6 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-5kmr6 ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-2t6js" podUID="c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.720693 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" event={"ID":"c1aa7752-d0d1-4dd9-9a49-8aee61349f10","Type":"ContainerDied","Data":"fdc0c275d44680f34c5896db231dc6fb365119fe362f7298cadde1125217c7fb"} Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.720744 4829 scope.go:117] "RemoveContainer" containerID="c7756e81f5c0bc6745c8a8fcf9502fd3e1057a68615797c8417cb9d226fc49eb" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.720846 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-dmrvn" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.721782 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-m6p2d"] Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.729977 4829 generic.go:334] "Generic (PLEG): container finished" podID="5919e153-3626-4764-8110-ae849b21ff17" containerID="66fd2865c1c504ddac447ea792768342a37d7a32f1d50b45059da5b204438574" exitCode=0 Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.730060 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kj4jn" event={"ID":"5919e153-3626-4764-8110-ae849b21ff17","Type":"ContainerDied","Data":"66fd2865c1c504ddac447ea792768342a37d7a32f1d50b45059da5b204438574"} Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.730161 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.734639 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m6p2d"] Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.769778 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dmrvn"] Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.779800 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-dmrvn"] Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.821989 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-scripts\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822036 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-ring-data-devices\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822092 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-combined-ca-bundle\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822147 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-dispersionconf\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822171 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-ring-data-devices\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822190 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-scripts\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822212 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-etc-swift\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822369 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-combined-ca-bundle\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822447 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kmr6\" (UniqueName: \"kubernetes.io/projected/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-kube-api-access-5kmr6\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822545 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5063471-4946-4763-9338-1ad3fbb4b9ef-etc-swift\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822572 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-dispersionconf\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822593 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-swiftconf\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822651 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-swiftconf\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.822680 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw7v4\" (UniqueName: \"kubernetes.io/projected/b5063471-4946-4763-9338-1ad3fbb4b9ef-kube-api-access-dw7v4\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.923822 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-swiftconf\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.923894 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw7v4\" (UniqueName: \"kubernetes.io/projected/b5063471-4946-4763-9338-1ad3fbb4b9ef-kube-api-access-dw7v4\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.923955 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-scripts\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.923991 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-ring-data-devices\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924045 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-combined-ca-bundle\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924095 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-dispersionconf\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924134 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-ring-data-devices\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924162 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-scripts\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924205 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-etc-swift\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924257 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-combined-ca-bundle\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924282 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kmr6\" (UniqueName: \"kubernetes.io/projected/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-kube-api-access-5kmr6\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924502 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5063471-4946-4763-9338-1ad3fbb4b9ef-etc-swift\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924554 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-dispersionconf\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924588 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-swiftconf\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924765 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-ring-data-devices\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.924807 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-scripts\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.925058 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-ring-data-devices\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.925309 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-scripts\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.925729 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5063471-4946-4763-9338-1ad3fbb4b9ef-etc-swift\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.926369 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-etc-swift\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.929829 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-swiftconf\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.930244 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-swiftconf\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.931021 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-dispersionconf\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.931058 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-combined-ca-bundle\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.931331 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-dispersionconf\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.934864 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-combined-ca-bundle\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.955877 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw7v4\" (UniqueName: \"kubernetes.io/projected/b5063471-4946-4763-9338-1ad3fbb4b9ef-kube-api-access-dw7v4\") pod \"swift-ring-rebalance-m6p2d\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:21 crc kubenswrapper[4829]: I1002 07:33:21.958738 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kmr6\" (UniqueName: \"kubernetes.io/projected/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-kube-api-access-5kmr6\") pod \"swift-ring-rebalance-2t6js\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.053731 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.737052 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.750285 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.857620 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-etc-swift\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.857758 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-dispersionconf\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858034 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858171 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kmr6\" (UniqueName: \"kubernetes.io/projected/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-kube-api-access-5kmr6\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858213 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-ring-data-devices\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858293 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-swiftconf\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858325 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-scripts\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858383 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-combined-ca-bundle\") pod \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\" (UID: \"c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225\") " Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858785 4829 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.858986 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.859518 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-scripts" (OuterVolumeSpecName: "scripts") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.861971 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-kube-api-access-5kmr6" (OuterVolumeSpecName: "kube-api-access-5kmr6") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "kube-api-access-5kmr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.862561 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.865309 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.865386 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" (UID: "c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.960356 4829 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.960386 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.960395 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.960404 4829 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.960413 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kmr6\" (UniqueName: \"kubernetes.io/projected/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-kube-api-access-5kmr6\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:22 crc kubenswrapper[4829]: I1002 07:33:22.960422 4829 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:23 crc kubenswrapper[4829]: I1002 07:33:23.473816 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1aa7752-d0d1-4dd9-9a49-8aee61349f10" path="/var/lib/kubelet/pods/c1aa7752-d0d1-4dd9-9a49-8aee61349f10/volumes" Oct 02 07:33:23 crc kubenswrapper[4829]: I1002 07:33:23.745180 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-2t6js" Oct 02 07:33:23 crc kubenswrapper[4829]: I1002 07:33:23.784903 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-2t6js"] Oct 02 07:33:23 crc kubenswrapper[4829]: I1002 07:33:23.790465 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-2t6js"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.361637 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-6wckj"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.366775 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.369586 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6wckj"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.489265 4829 scope.go:117] "RemoveContainer" containerID="33c62eb6e94f2adc8f0e0863e6aab1d0c33debbca0d6bc169b4a8c7b96eabc63" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.495872 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp8wd\" (UniqueName: \"kubernetes.io/projected/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42-kube-api-access-zp8wd\") pod \"keystone-db-create-6wckj\" (UID: \"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42\") " pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.609598 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp8wd\" (UniqueName: \"kubernetes.io/projected/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42-kube-api-access-zp8wd\") pod \"keystone-db-create-6wckj\" (UID: \"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42\") " pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.618613 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-lh5bj"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.620184 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.627438 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp8wd\" (UniqueName: \"kubernetes.io/projected/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42-kube-api-access-zp8wd\") pod \"keystone-db-create-6wckj\" (UID: \"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42\") " pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.634737 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-lh5bj"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.690461 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.812193 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rmt9\" (UniqueName: \"kubernetes.io/projected/24271dcb-9076-48b4-9698-a3e97a4bd625-kube-api-access-6rmt9\") pod \"placement-db-create-lh5bj\" (UID: \"24271dcb-9076-48b4-9698-a3e97a4bd625\") " pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.861764 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-lvvhk"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.863161 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.869926 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lvvhk"] Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.914717 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rmt9\" (UniqueName: \"kubernetes.io/projected/24271dcb-9076-48b4-9698-a3e97a4bd625-kube-api-access-6rmt9\") pod \"placement-db-create-lh5bj\" (UID: \"24271dcb-9076-48b4-9698-a3e97a4bd625\") " pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.939464 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rmt9\" (UniqueName: \"kubernetes.io/projected/24271dcb-9076-48b4-9698-a3e97a4bd625-kube-api-access-6rmt9\") pod \"placement-db-create-lh5bj\" (UID: \"24271dcb-9076-48b4-9698-a3e97a4bd625\") " pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:24 crc kubenswrapper[4829]: I1002 07:33:24.947782 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-m6p2d"] Oct 02 07:33:24 crc kubenswrapper[4829]: W1002 07:33:24.953460 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5063471_4946_4763_9338_1ad3fbb4b9ef.slice/crio-d608ca387bb1d5a254ad8d7a064225bcb5d407e9599f44a5eaf0d74e4028f6f7 WatchSource:0}: Error finding container d608ca387bb1d5a254ad8d7a064225bcb5d407e9599f44a5eaf0d74e4028f6f7: Status 404 returned error can't find the container with id d608ca387bb1d5a254ad8d7a064225bcb5d407e9599f44a5eaf0d74e4028f6f7 Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.016703 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glzbk\" (UniqueName: \"kubernetes.io/projected/9059dfc6-81c9-4224-b364-94d9d2f50d67-kube-api-access-glzbk\") pod \"glance-db-create-lvvhk\" (UID: \"9059dfc6-81c9-4224-b364-94d9d2f50d67\") " pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.027149 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.122718 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glzbk\" (UniqueName: \"kubernetes.io/projected/9059dfc6-81c9-4224-b364-94d9d2f50d67-kube-api-access-glzbk\") pod \"glance-db-create-lvvhk\" (UID: \"9059dfc6-81c9-4224-b364-94d9d2f50d67\") " pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.139840 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-6wckj"] Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.147118 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glzbk\" (UniqueName: \"kubernetes.io/projected/9059dfc6-81c9-4224-b364-94d9d2f50d67-kube-api-access-glzbk\") pod \"glance-db-create-lvvhk\" (UID: \"9059dfc6-81c9-4224-b364-94d9d2f50d67\") " pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:25 crc kubenswrapper[4829]: W1002 07:33:25.152445 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc14ab4eb_2e4c_4196_8689_c58e5c6e4a42.slice/crio-a63cbd440f4e972db06945e8cc42d864a2a57576ad03ef73a313f9174812a5ff WatchSource:0}: Error finding container a63cbd440f4e972db06945e8cc42d864a2a57576ad03ef73a313f9174812a5ff: Status 404 returned error can't find the container with id a63cbd440f4e972db06945e8cc42d864a2a57576ad03ef73a313f9174812a5ff Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.214817 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.484022 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225" path="/var/lib/kubelet/pods/c8c5f5fe-e6b5-40fd-81fa-cbe0ae7f8225/volumes" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.485513 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-lh5bj"] Oct 02 07:33:25 crc kubenswrapper[4829]: W1002 07:33:25.493209 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24271dcb_9076_48b4_9698_a3e97a4bd625.slice/crio-8ad453931001dd7f5294a45dd4ec9fea2ad01a315dd15b9b4031b018cd46bf61 WatchSource:0}: Error finding container 8ad453931001dd7f5294a45dd4ec9fea2ad01a315dd15b9b4031b018cd46bf61: Status 404 returned error can't find the container with id 8ad453931001dd7f5294a45dd4ec9fea2ad01a315dd15b9b4031b018cd46bf61 Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.530415 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:25 crc kubenswrapper[4829]: E1002 07:33:25.530706 4829 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 07:33:25 crc kubenswrapper[4829]: E1002 07:33:25.530774 4829 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 07:33:25 crc kubenswrapper[4829]: E1002 07:33:25.530900 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift podName:9da44bca-624e-49ca-8fd2-e0b0974c9ae5 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:33.530876481 +0000 UTC m=+1004.870524886 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift") pod "swift-storage-0" (UID: "9da44bca-624e-49ca-8fd2-e0b0974c9ae5") : configmap "swift-ring-files" not found Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.615101 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-lvvhk"] Oct 02 07:33:25 crc kubenswrapper[4829]: W1002 07:33:25.628019 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9059dfc6_81c9_4224_b364_94d9d2f50d67.slice/crio-c6251e788c3b5c8003b24a07de4aeba0f8c9fdcc42d1d93469295a2765800e9c WatchSource:0}: Error finding container c6251e788c3b5c8003b24a07de4aeba0f8c9fdcc42d1d93469295a2765800e9c: Status 404 returned error can't find the container with id c6251e788c3b5c8003b24a07de4aeba0f8c9fdcc42d1d93469295a2765800e9c Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.766616 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerStarted","Data":"4fd0cc46b9bf567752d33e830edff9203b00336e0a8ea83b189d5c929c40395a"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.768340 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lvvhk" event={"ID":"9059dfc6-81c9-4224-b364-94d9d2f50d67","Type":"ContainerStarted","Data":"c6251e788c3b5c8003b24a07de4aeba0f8c9fdcc42d1d93469295a2765800e9c"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.770392 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lh5bj" event={"ID":"24271dcb-9076-48b4-9698-a3e97a4bd625","Type":"ContainerStarted","Data":"1af76466372c9c037930e400e2270fe1e220f18d8c50a01d7806b3f450cf6669"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.770486 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lh5bj" event={"ID":"24271dcb-9076-48b4-9698-a3e97a4bd625","Type":"ContainerStarted","Data":"8ad453931001dd7f5294a45dd4ec9fea2ad01a315dd15b9b4031b018cd46bf61"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.772030 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kj4jn" event={"ID":"5919e153-3626-4764-8110-ae849b21ff17","Type":"ContainerStarted","Data":"6f4ebf4a95d5f6ba9071dd198b392585ee5ec237cba3ceb3d9ba50b2058a7128"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.772714 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.773944 4829 generic.go:334] "Generic (PLEG): container finished" podID="c14ab4eb-2e4c-4196-8689-c58e5c6e4a42" containerID="15af294706718995e7df334a59da66bc1211326e23ae46f9db15f14da5ce4311" exitCode=0 Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.773973 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6wckj" event={"ID":"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42","Type":"ContainerDied","Data":"15af294706718995e7df334a59da66bc1211326e23ae46f9db15f14da5ce4311"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.774049 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6wckj" event={"ID":"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42","Type":"ContainerStarted","Data":"a63cbd440f4e972db06945e8cc42d864a2a57576ad03ef73a313f9174812a5ff"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.774859 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6p2d" event={"ID":"b5063471-4946-4763-9338-1ad3fbb4b9ef","Type":"ContainerStarted","Data":"d608ca387bb1d5a254ad8d7a064225bcb5d407e9599f44a5eaf0d74e4028f6f7"} Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.793265 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-lh5bj" podStartSLOduration=1.7932424980000001 podStartE2EDuration="1.793242498s" podCreationTimestamp="2025-10-02 07:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:25.784983672 +0000 UTC m=+997.124632087" watchObservedRunningTime="2025-10-02 07:33:25.793242498 +0000 UTC m=+997.132890903" Oct 02 07:33:25 crc kubenswrapper[4829]: I1002 07:33:25.809772 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-kj4jn" podStartSLOduration=9.809752783 podStartE2EDuration="9.809752783s" podCreationTimestamp="2025-10-02 07:33:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:25.804566004 +0000 UTC m=+997.144214419" watchObservedRunningTime="2025-10-02 07:33:25.809752783 +0000 UTC m=+997.149401188" Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.610508 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-kc6lg"] Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.612068 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.623400 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-kc6lg"] Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.744286 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.749775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5cjl\" (UniqueName: \"kubernetes.io/projected/38a3be44-070a-4d1b-acec-d3d7baac2615-kube-api-access-x5cjl\") pod \"watcher-db-create-kc6lg\" (UID: \"38a3be44-070a-4d1b-acec-d3d7baac2615\") " pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.786307 4829 generic.go:334] "Generic (PLEG): container finished" podID="9059dfc6-81c9-4224-b364-94d9d2f50d67" containerID="ed6cef3818f9f4fcdbee3cf10d168ef0310b0257286cf79c3d92c6bbe8663bbd" exitCode=0 Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.786373 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lvvhk" event={"ID":"9059dfc6-81c9-4224-b364-94d9d2f50d67","Type":"ContainerDied","Data":"ed6cef3818f9f4fcdbee3cf10d168ef0310b0257286cf79c3d92c6bbe8663bbd"} Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.788941 4829 generic.go:334] "Generic (PLEG): container finished" podID="24271dcb-9076-48b4-9698-a3e97a4bd625" containerID="1af76466372c9c037930e400e2270fe1e220f18d8c50a01d7806b3f450cf6669" exitCode=0 Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.794296 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lh5bj" event={"ID":"24271dcb-9076-48b4-9698-a3e97a4bd625","Type":"ContainerDied","Data":"1af76466372c9c037930e400e2270fe1e220f18d8c50a01d7806b3f450cf6669"} Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.851503 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5cjl\" (UniqueName: \"kubernetes.io/projected/38a3be44-070a-4d1b-acec-d3d7baac2615-kube-api-access-x5cjl\") pod \"watcher-db-create-kc6lg\" (UID: \"38a3be44-070a-4d1b-acec-d3d7baac2615\") " pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:26 crc kubenswrapper[4829]: I1002 07:33:26.996058 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5cjl\" (UniqueName: \"kubernetes.io/projected/38a3be44-070a-4d1b-acec-d3d7baac2615-kube-api-access-x5cjl\") pod \"watcher-db-create-kc6lg\" (UID: \"38a3be44-070a-4d1b-acec-d3d7baac2615\") " pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:27 crc kubenswrapper[4829]: I1002 07:33:27.240062 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:27 crc kubenswrapper[4829]: I1002 07:33:27.809144 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerStarted","Data":"9c33059abbde95491ae84e5c6e8fb89d301e4b1d50816cd2a36e60454d72a0c8"} Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.306382 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.320336 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.341923 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.385556 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glzbk\" (UniqueName: \"kubernetes.io/projected/9059dfc6-81c9-4224-b364-94d9d2f50d67-kube-api-access-glzbk\") pod \"9059dfc6-81c9-4224-b364-94d9d2f50d67\" (UID: \"9059dfc6-81c9-4224-b364-94d9d2f50d67\") " Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.385738 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rmt9\" (UniqueName: \"kubernetes.io/projected/24271dcb-9076-48b4-9698-a3e97a4bd625-kube-api-access-6rmt9\") pod \"24271dcb-9076-48b4-9698-a3e97a4bd625\" (UID: \"24271dcb-9076-48b4-9698-a3e97a4bd625\") " Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.394270 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24271dcb-9076-48b4-9698-a3e97a4bd625-kube-api-access-6rmt9" (OuterVolumeSpecName: "kube-api-access-6rmt9") pod "24271dcb-9076-48b4-9698-a3e97a4bd625" (UID: "24271dcb-9076-48b4-9698-a3e97a4bd625"). InnerVolumeSpecName "kube-api-access-6rmt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.394965 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9059dfc6-81c9-4224-b364-94d9d2f50d67-kube-api-access-glzbk" (OuterVolumeSpecName: "kube-api-access-glzbk") pod "9059dfc6-81c9-4224-b364-94d9d2f50d67" (UID: "9059dfc6-81c9-4224-b364-94d9d2f50d67"). InnerVolumeSpecName "kube-api-access-glzbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.487652 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp8wd\" (UniqueName: \"kubernetes.io/projected/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42-kube-api-access-zp8wd\") pod \"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42\" (UID: \"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42\") " Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.488415 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rmt9\" (UniqueName: \"kubernetes.io/projected/24271dcb-9076-48b4-9698-a3e97a4bd625-kube-api-access-6rmt9\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.488444 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glzbk\" (UniqueName: \"kubernetes.io/projected/9059dfc6-81c9-4224-b364-94d9d2f50d67-kube-api-access-glzbk\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.490701 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42-kube-api-access-zp8wd" (OuterVolumeSpecName: "kube-api-access-zp8wd") pod "c14ab4eb-2e4c-4196-8689-c58e5c6e4a42" (UID: "c14ab4eb-2e4c-4196-8689-c58e5c6e4a42"). InnerVolumeSpecName "kube-api-access-zp8wd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.591350 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp8wd\" (UniqueName: \"kubernetes.io/projected/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42-kube-api-access-zp8wd\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.700874 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-kc6lg"] Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.822153 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-6wckj" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.822174 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-6wckj" event={"ID":"c14ab4eb-2e4c-4196-8689-c58e5c6e4a42","Type":"ContainerDied","Data":"a63cbd440f4e972db06945e8cc42d864a2a57576ad03ef73a313f9174812a5ff"} Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.822353 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a63cbd440f4e972db06945e8cc42d864a2a57576ad03ef73a313f9174812a5ff" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.825507 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-lvvhk" event={"ID":"9059dfc6-81c9-4224-b364-94d9d2f50d67","Type":"ContainerDied","Data":"c6251e788c3b5c8003b24a07de4aeba0f8c9fdcc42d1d93469295a2765800e9c"} Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.825564 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6251e788c3b5c8003b24a07de4aeba0f8c9fdcc42d1d93469295a2765800e9c" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.825634 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-lvvhk" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.830741 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-lh5bj" event={"ID":"24271dcb-9076-48b4-9698-a3e97a4bd625","Type":"ContainerDied","Data":"8ad453931001dd7f5294a45dd4ec9fea2ad01a315dd15b9b4031b018cd46bf61"} Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.830825 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ad453931001dd7f5294a45dd4ec9fea2ad01a315dd15b9b4031b018cd46bf61" Oct 02 07:33:28 crc kubenswrapper[4829]: I1002 07:33:28.830875 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-lh5bj" Oct 02 07:33:28 crc kubenswrapper[4829]: W1002 07:33:28.991315 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38a3be44_070a_4d1b_acec_d3d7baac2615.slice/crio-71e064289af18339241fb974bef2d32ecf2d941f183f2844ac33d018d2296864 WatchSource:0}: Error finding container 71e064289af18339241fb974bef2d32ecf2d941f183f2844ac33d018d2296864: Status 404 returned error can't find the container with id 71e064289af18339241fb974bef2d32ecf2d941f183f2844ac33d018d2296864 Oct 02 07:33:29 crc kubenswrapper[4829]: I1002 07:33:29.845209 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6p2d" event={"ID":"b5063471-4946-4763-9338-1ad3fbb4b9ef","Type":"ContainerStarted","Data":"6e2592b71fdd8738b3957ffdb0970a11344b747798f342fa83ac332718df03c7"} Oct 02 07:33:29 crc kubenswrapper[4829]: I1002 07:33:29.850108 4829 generic.go:334] "Generic (PLEG): container finished" podID="38a3be44-070a-4d1b-acec-d3d7baac2615" containerID="2a9b9f8b92b38426762f0678d8ba976f47209d1139aa56ea90af2247985aece8" exitCode=0 Oct 02 07:33:29 crc kubenswrapper[4829]: I1002 07:33:29.850193 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-kc6lg" event={"ID":"38a3be44-070a-4d1b-acec-d3d7baac2615","Type":"ContainerDied","Data":"2a9b9f8b92b38426762f0678d8ba976f47209d1139aa56ea90af2247985aece8"} Oct 02 07:33:29 crc kubenswrapper[4829]: I1002 07:33:29.850271 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-kc6lg" event={"ID":"38a3be44-070a-4d1b-acec-d3d7baac2615","Type":"ContainerStarted","Data":"71e064289af18339241fb974bef2d32ecf2d941f183f2844ac33d018d2296864"} Oct 02 07:33:29 crc kubenswrapper[4829]: I1002 07:33:29.871968 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-m6p2d" podStartSLOduration=4.802817169 podStartE2EDuration="8.871942257s" podCreationTimestamp="2025-10-02 07:33:21 +0000 UTC" firstStartedPulling="2025-10-02 07:33:24.954885589 +0000 UTC m=+996.294533994" lastFinishedPulling="2025-10-02 07:33:29.024010667 +0000 UTC m=+1000.363659082" observedRunningTime="2025-10-02 07:33:29.863847246 +0000 UTC m=+1001.203495721" watchObservedRunningTime="2025-10-02 07:33:29.871942257 +0000 UTC m=+1001.211590672" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.220128 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.351993 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5cjl\" (UniqueName: \"kubernetes.io/projected/38a3be44-070a-4d1b-acec-d3d7baac2615-kube-api-access-x5cjl\") pod \"38a3be44-070a-4d1b-acec-d3d7baac2615\" (UID: \"38a3be44-070a-4d1b-acec-d3d7baac2615\") " Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.359140 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38a3be44-070a-4d1b-acec-d3d7baac2615-kube-api-access-x5cjl" (OuterVolumeSpecName: "kube-api-access-x5cjl") pod "38a3be44-070a-4d1b-acec-d3d7baac2615" (UID: "38a3be44-070a-4d1b-acec-d3d7baac2615"). InnerVolumeSpecName "kube-api-access-x5cjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.455167 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5cjl\" (UniqueName: \"kubernetes.io/projected/38a3be44-070a-4d1b-acec-d3d7baac2615-kube-api-access-x5cjl\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.870643 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerStarted","Data":"bcd7a8bf4486d27e6f4d8427b9c5722ebfeb26a8149952713aa27fcdb806e982"} Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.872768 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-kc6lg" event={"ID":"38a3be44-070a-4d1b-acec-d3d7baac2615","Type":"ContainerDied","Data":"71e064289af18339241fb974bef2d32ecf2d941f183f2844ac33d018d2296864"} Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.872980 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71e064289af18339241fb974bef2d32ecf2d941f183f2844ac33d018d2296864" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.873043 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-kc6lg" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.910418 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.911957 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=8.398294616 podStartE2EDuration="45.911933207s" podCreationTimestamp="2025-10-02 07:32:46 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.239352767 +0000 UTC m=+964.579001172" lastFinishedPulling="2025-10-02 07:33:30.752991358 +0000 UTC m=+1002.092639763" observedRunningTime="2025-10-02 07:33:31.910612565 +0000 UTC m=+1003.250261060" watchObservedRunningTime="2025-10-02 07:33:31.911933207 +0000 UTC m=+1003.251581622" Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.976739 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fjdw2"] Oct 02 07:33:31 crc kubenswrapper[4829]: I1002 07:33:31.977002 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" containerName="dnsmasq-dns" containerID="cri-o://f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6" gracePeriod=10 Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.423413 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.572918 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-dns-svc\") pod \"9c93fa9a-4673-46f4-b432-ff794f518510\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.573111 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-config\") pod \"9c93fa9a-4673-46f4-b432-ff794f518510\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.573219 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmq5n\" (UniqueName: \"kubernetes.io/projected/9c93fa9a-4673-46f4-b432-ff794f518510-kube-api-access-jmq5n\") pod \"9c93fa9a-4673-46f4-b432-ff794f518510\" (UID: \"9c93fa9a-4673-46f4-b432-ff794f518510\") " Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.610451 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c93fa9a-4673-46f4-b432-ff794f518510-kube-api-access-jmq5n" (OuterVolumeSpecName: "kube-api-access-jmq5n") pod "9c93fa9a-4673-46f4-b432-ff794f518510" (UID: "9c93fa9a-4673-46f4-b432-ff794f518510"). InnerVolumeSpecName "kube-api-access-jmq5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.675491 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmq5n\" (UniqueName: \"kubernetes.io/projected/9c93fa9a-4673-46f4-b432-ff794f518510-kube-api-access-jmq5n\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.677844 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-config" (OuterVolumeSpecName: "config") pod "9c93fa9a-4673-46f4-b432-ff794f518510" (UID: "9c93fa9a-4673-46f4-b432-ff794f518510"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.687011 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9c93fa9a-4673-46f4-b432-ff794f518510" (UID: "9c93fa9a-4673-46f4-b432-ff794f518510"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.776926 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.776961 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9c93fa9a-4673-46f4-b432-ff794f518510-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.830253 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.863899 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.866886 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.881933 4829 generic.go:334] "Generic (PLEG): container finished" podID="9c93fa9a-4673-46f4-b432-ff794f518510" containerID="f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6" exitCode=0 Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.885182 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.886399 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" event={"ID":"9c93fa9a-4673-46f4-b432-ff794f518510","Type":"ContainerDied","Data":"f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6"} Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.886447 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fjdw2" event={"ID":"9c93fa9a-4673-46f4-b432-ff794f518510","Type":"ContainerDied","Data":"e77f5e2bb4c31856d08cfb03a917fa61a5b0cba6f62542c7f82915c07a7ba178"} Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.886466 4829 scope.go:117] "RemoveContainer" containerID="f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.887012 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.907102 4829 scope.go:117] "RemoveContainer" containerID="2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.955090 4829 scope.go:117] "RemoveContainer" containerID="f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6" Oct 02 07:33:32 crc kubenswrapper[4829]: E1002 07:33:32.965012 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6\": container with ID starting with f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6 not found: ID does not exist" containerID="f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.968734 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6"} err="failed to get container status \"f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6\": rpc error: code = NotFound desc = could not find container \"f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6\": container with ID starting with f8ed576ed7540335052cc30da0b1025227ee915e00f27f7130cf82311f74e5b6 not found: ID does not exist" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.968875 4829 scope.go:117] "RemoveContainer" containerID="2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918" Oct 02 07:33:32 crc kubenswrapper[4829]: E1002 07:33:32.969359 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918\": container with ID starting with 2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918 not found: ID does not exist" containerID="2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.969386 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918"} err="failed to get container status \"2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918\": rpc error: code = NotFound desc = could not find container \"2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918\": container with ID starting with 2fd5f886faa7781305633ad487a446c02821889d59b28a1e3f5ee0df1736a918 not found: ID does not exist" Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.982364 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fjdw2"] Oct 02 07:33:32 crc kubenswrapper[4829]: I1002 07:33:32.986848 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fjdw2"] Oct 02 07:33:33 crc kubenswrapper[4829]: I1002 07:33:33.477979 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" path="/var/lib/kubelet/pods/9c93fa9a-4673-46f4-b432-ff794f518510/volumes" Oct 02 07:33:33 crc kubenswrapper[4829]: I1002 07:33:33.609861 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:33 crc kubenswrapper[4829]: E1002 07:33:33.610221 4829 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 07:33:33 crc kubenswrapper[4829]: E1002 07:33:33.610306 4829 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 07:33:33 crc kubenswrapper[4829]: E1002 07:33:33.610454 4829 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift podName:9da44bca-624e-49ca-8fd2-e0b0974c9ae5 nodeName:}" failed. No retries permitted until 2025-10-02 07:33:49.610377371 +0000 UTC m=+1020.950025826 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift") pod "swift-storage-0" (UID: "9da44bca-624e-49ca-8fd2-e0b0974c9ae5") : configmap "swift-ring-files" not found Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.711857 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3b4a-account-create-w6wq5"] Oct 02 07:33:34 crc kubenswrapper[4829]: E1002 07:33:34.712175 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c14ab4eb-2e4c-4196-8689-c58e5c6e4a42" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712186 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c14ab4eb-2e4c-4196-8689-c58e5c6e4a42" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: E1002 07:33:34.712197 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" containerName="dnsmasq-dns" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712203 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" containerName="dnsmasq-dns" Oct 02 07:33:34 crc kubenswrapper[4829]: E1002 07:33:34.712214 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9059dfc6-81c9-4224-b364-94d9d2f50d67" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712234 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9059dfc6-81c9-4224-b364-94d9d2f50d67" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: E1002 07:33:34.712250 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" containerName="init" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712256 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" containerName="init" Oct 02 07:33:34 crc kubenswrapper[4829]: E1002 07:33:34.712266 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38a3be44-070a-4d1b-acec-d3d7baac2615" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712272 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="38a3be44-070a-4d1b-acec-d3d7baac2615" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: E1002 07:33:34.712281 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24271dcb-9076-48b4-9698-a3e97a4bd625" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712287 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="24271dcb-9076-48b4-9698-a3e97a4bd625" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712437 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9059dfc6-81c9-4224-b364-94d9d2f50d67" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712455 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="24271dcb-9076-48b4-9698-a3e97a4bd625" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712473 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="38a3be44-070a-4d1b-acec-d3d7baac2615" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712484 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c14ab4eb-2e4c-4196-8689-c58e5c6e4a42" containerName="mariadb-database-create" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.712492 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c93fa9a-4673-46f4-b432-ff794f518510" containerName="dnsmasq-dns" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.713025 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.714925 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.729447 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3b4a-account-create-w6wq5"] Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.831302 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2xth\" (UniqueName: \"kubernetes.io/projected/ac2d616a-91bd-4aa2-b818-09bd73c59320-kube-api-access-m2xth\") pod \"placement-3b4a-account-create-w6wq5\" (UID: \"ac2d616a-91bd-4aa2-b818-09bd73c59320\") " pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.932354 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2xth\" (UniqueName: \"kubernetes.io/projected/ac2d616a-91bd-4aa2-b818-09bd73c59320-kube-api-access-m2xth\") pod \"placement-3b4a-account-create-w6wq5\" (UID: \"ac2d616a-91bd-4aa2-b818-09bd73c59320\") " pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:34 crc kubenswrapper[4829]: I1002 07:33:34.950146 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2xth\" (UniqueName: \"kubernetes.io/projected/ac2d616a-91bd-4aa2-b818-09bd73c59320-kube-api-access-m2xth\") pod \"placement-3b4a-account-create-w6wq5\" (UID: \"ac2d616a-91bd-4aa2-b818-09bd73c59320\") " pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.018127 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-4939-account-create-2wx9l"] Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.019052 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.021781 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.032668 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4939-account-create-2wx9l"] Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.037294 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.134445 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz82h\" (UniqueName: \"kubernetes.io/projected/9ae65aa7-5414-49a2-bd11-dd2209873f7e-kube-api-access-hz82h\") pod \"glance-4939-account-create-2wx9l\" (UID: \"9ae65aa7-5414-49a2-bd11-dd2209873f7e\") " pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.238762 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.239476 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz82h\" (UniqueName: \"kubernetes.io/projected/9ae65aa7-5414-49a2-bd11-dd2209873f7e-kube-api-access-hz82h\") pod \"glance-4939-account-create-2wx9l\" (UID: \"9ae65aa7-5414-49a2-bd11-dd2209873f7e\") " pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.282311 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz82h\" (UniqueName: \"kubernetes.io/projected/9ae65aa7-5414-49a2-bd11-dd2209873f7e-kube-api-access-hz82h\") pod \"glance-4939-account-create-2wx9l\" (UID: \"9ae65aa7-5414-49a2-bd11-dd2209873f7e\") " pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.417153 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.488663 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3b4a-account-create-w6wq5"] Oct 02 07:33:35 crc kubenswrapper[4829]: W1002 07:33:35.490124 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac2d616a_91bd_4aa2_b818_09bd73c59320.slice/crio-b8e027559038162e1a32d598a7fa0eae05e3a22e5a09873cd59405298d3f6613 WatchSource:0}: Error finding container b8e027559038162e1a32d598a7fa0eae05e3a22e5a09873cd59405298d3f6613: Status 404 returned error can't find the container with id b8e027559038162e1a32d598a7fa0eae05e3a22e5a09873cd59405298d3f6613 Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.584904 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-njw55" podUID="91ff1e5f-73e2-4237-b201-84065a586553" containerName="ovn-controller" probeResult="failure" output=< Oct 02 07:33:35 crc kubenswrapper[4829]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 07:33:35 crc kubenswrapper[4829]: > Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.887096 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4939-account-create-2wx9l"] Oct 02 07:33:35 crc kubenswrapper[4829]: W1002 07:33:35.890266 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ae65aa7_5414_49a2_bd11_dd2209873f7e.slice/crio-308b8c2ee2df531bec333613c7b7e09975d723127bc469ba469e10f6d9be9998 WatchSource:0}: Error finding container 308b8c2ee2df531bec333613c7b7e09975d723127bc469ba469e10f6d9be9998: Status 404 returned error can't find the container with id 308b8c2ee2df531bec333613c7b7e09975d723127bc469ba469e10f6d9be9998 Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.921481 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4939-account-create-2wx9l" event={"ID":"9ae65aa7-5414-49a2-bd11-dd2209873f7e","Type":"ContainerStarted","Data":"308b8c2ee2df531bec333613c7b7e09975d723127bc469ba469e10f6d9be9998"} Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.923560 4829 generic.go:334] "Generic (PLEG): container finished" podID="ac2d616a-91bd-4aa2-b818-09bd73c59320" containerID="83f4ac4c30cc9d4bbb7770e1caa7034916ee286611cc93005538f68abe316fe2" exitCode=0 Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.923643 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3b4a-account-create-w6wq5" event={"ID":"ac2d616a-91bd-4aa2-b818-09bd73c59320","Type":"ContainerDied","Data":"83f4ac4c30cc9d4bbb7770e1caa7034916ee286611cc93005538f68abe316fe2"} Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.923714 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3b4a-account-create-w6wq5" event={"ID":"ac2d616a-91bd-4aa2-b818-09bd73c59320","Type":"ContainerStarted","Data":"b8e027559038162e1a32d598a7fa0eae05e3a22e5a09873cd59405298d3f6613"} Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.923952 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="prometheus" containerID="cri-o://4fd0cc46b9bf567752d33e830edff9203b00336e0a8ea83b189d5c929c40395a" gracePeriod=600 Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.924101 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="thanos-sidecar" containerID="cri-o://bcd7a8bf4486d27e6f4d8427b9c5722ebfeb26a8149952713aa27fcdb806e982" gracePeriod=600 Oct 02 07:33:35 crc kubenswrapper[4829]: I1002 07:33:35.924166 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="config-reloader" containerID="cri-o://9c33059abbde95491ae84e5c6e8fb89d301e4b1d50816cd2a36e60454d72a0c8" gracePeriod=600 Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.934468 4829 generic.go:334] "Generic (PLEG): container finished" podID="9ae65aa7-5414-49a2-bd11-dd2209873f7e" containerID="34a54c0c6302c4c359abf36f5599e8ea69273606ea5a3310f8eeca7f4eb10b1d" exitCode=0 Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.934525 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4939-account-create-2wx9l" event={"ID":"9ae65aa7-5414-49a2-bd11-dd2209873f7e","Type":"ContainerDied","Data":"34a54c0c6302c4c359abf36f5599e8ea69273606ea5a3310f8eeca7f4eb10b1d"} Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.937488 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6p2d" event={"ID":"b5063471-4946-4763-9338-1ad3fbb4b9ef","Type":"ContainerDied","Data":"6e2592b71fdd8738b3957ffdb0970a11344b747798f342fa83ac332718df03c7"} Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.937518 4829 generic.go:334] "Generic (PLEG): container finished" podID="b5063471-4946-4763-9338-1ad3fbb4b9ef" containerID="6e2592b71fdd8738b3957ffdb0970a11344b747798f342fa83ac332718df03c7" exitCode=0 Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951273 4829 generic.go:334] "Generic (PLEG): container finished" podID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerID="bcd7a8bf4486d27e6f4d8427b9c5722ebfeb26a8149952713aa27fcdb806e982" exitCode=0 Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951312 4829 generic.go:334] "Generic (PLEG): container finished" podID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerID="9c33059abbde95491ae84e5c6e8fb89d301e4b1d50816cd2a36e60454d72a0c8" exitCode=0 Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951326 4829 generic.go:334] "Generic (PLEG): container finished" podID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerID="4fd0cc46b9bf567752d33e830edff9203b00336e0a8ea83b189d5c929c40395a" exitCode=0 Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951477 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerDied","Data":"bcd7a8bf4486d27e6f4d8427b9c5722ebfeb26a8149952713aa27fcdb806e982"} Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951506 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerDied","Data":"9c33059abbde95491ae84e5c6e8fb89d301e4b1d50816cd2a36e60454d72a0c8"} Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951520 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerDied","Data":"4fd0cc46b9bf567752d33e830edff9203b00336e0a8ea83b189d5c929c40395a"} Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951532 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"61f93d4a-ce37-43ad-85d0-4be548d45c3b","Type":"ContainerDied","Data":"ec5d5816bb3bb652b9384fb03f5b6bd8c18011da7d3a56b6e1e2380a796dbf0c"} Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.951545 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec5d5816bb3bb652b9384fb03f5b6bd8c18011da7d3a56b6e1e2380a796dbf0c" Oct 02 07:33:36 crc kubenswrapper[4829]: I1002 07:33:36.960046 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.070530 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-web-config\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.070614 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/61f93d4a-ce37-43ad-85d0-4be548d45c3b-prometheus-metric-storage-rulefiles-0\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.070643 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-tls-assets\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.070791 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2sgw\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-kube-api-access-k2sgw\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.070991 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.071031 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-thanos-prometheus-http-client-file\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.071067 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.071092 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config-out\") pod \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\" (UID: \"61f93d4a-ce37-43ad-85d0-4be548d45c3b\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.071391 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61f93d4a-ce37-43ad-85d0-4be548d45c3b-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.071616 4829 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/61f93d4a-ce37-43ad-85d0-4be548d45c3b-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.079894 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config" (OuterVolumeSpecName: "config") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.080166 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-kube-api-access-k2sgw" (OuterVolumeSpecName: "kube-api-access-k2sgw") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "kube-api-access-k2sgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.082043 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.082445 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.084717 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config-out" (OuterVolumeSpecName: "config-out") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.098655 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "pvc-81c3850c-e595-4f65-b168-53c52190636d". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.109446 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-web-config" (OuterVolumeSpecName: "web-config") pod "61f93d4a-ce37-43ad-85d0-4be548d45c3b" (UID: "61f93d4a-ce37-43ad-85d0-4be548d45c3b"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175483 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2sgw\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-kube-api-access-k2sgw\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175537 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") on node \"crc\" " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175552 4829 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175566 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175578 4829 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/61f93d4a-ce37-43ad-85d0-4be548d45c3b-config-out\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175591 4829 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/61f93d4a-ce37-43ad-85d0-4be548d45c3b-web-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.175601 4829 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/61f93d4a-ce37-43ad-85d0-4be548d45c3b-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.202125 4829 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.202304 4829 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-81c3850c-e595-4f65-b168-53c52190636d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d") on node "crc" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.277674 4829 reconciler_common.go:293] "Volume detached for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.345725 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.480981 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2xth\" (UniqueName: \"kubernetes.io/projected/ac2d616a-91bd-4aa2-b818-09bd73c59320-kube-api-access-m2xth\") pod \"ac2d616a-91bd-4aa2-b818-09bd73c59320\" (UID: \"ac2d616a-91bd-4aa2-b818-09bd73c59320\") " Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.484458 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac2d616a-91bd-4aa2-b818-09bd73c59320-kube-api-access-m2xth" (OuterVolumeSpecName: "kube-api-access-m2xth") pod "ac2d616a-91bd-4aa2-b818-09bd73c59320" (UID: "ac2d616a-91bd-4aa2-b818-09bd73c59320"). InnerVolumeSpecName "kube-api-access-m2xth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.582871 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2xth\" (UniqueName: \"kubernetes.io/projected/ac2d616a-91bd-4aa2-b818-09bd73c59320-kube-api-access-m2xth\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.962969 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3b4a-account-create-w6wq5" event={"ID":"ac2d616a-91bd-4aa2-b818-09bd73c59320","Type":"ContainerDied","Data":"b8e027559038162e1a32d598a7fa0eae05e3a22e5a09873cd59405298d3f6613"} Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.963286 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.963302 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8e027559038162e1a32d598a7fa0eae05e3a22e5a09873cd59405298d3f6613" Oct 02 07:33:37 crc kubenswrapper[4829]: I1002 07:33:37.963044 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3b4a-account-create-w6wq5" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.010531 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.017574 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.050203 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:33:38 crc kubenswrapper[4829]: E1002 07:33:38.050759 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="prometheus" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.050790 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="prometheus" Oct 02 07:33:38 crc kubenswrapper[4829]: E1002 07:33:38.050821 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="thanos-sidecar" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.050831 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="thanos-sidecar" Oct 02 07:33:38 crc kubenswrapper[4829]: E1002 07:33:38.050853 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="config-reloader" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.050863 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="config-reloader" Oct 02 07:33:38 crc kubenswrapper[4829]: E1002 07:33:38.050880 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac2d616a-91bd-4aa2-b818-09bd73c59320" containerName="mariadb-account-create" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.050892 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac2d616a-91bd-4aa2-b818-09bd73c59320" containerName="mariadb-account-create" Oct 02 07:33:38 crc kubenswrapper[4829]: E1002 07:33:38.050906 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="init-config-reloader" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.050916 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="init-config-reloader" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.051200 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="prometheus" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.051273 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="thanos-sidecar" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.051290 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac2d616a-91bd-4aa2-b818-09bd73c59320" containerName="mariadb-account-create" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.051306 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" containerName="config-reloader" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.053708 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.062273 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.063419 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.064357 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.067197 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hrzwj" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.067357 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.070145 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.085716 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.093257 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.201758 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.201797 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4469e851-270f-47ee-a523-e3fb56d8201f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.201823 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.201856 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.201875 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-config\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.202013 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.202070 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hnpl\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-kube-api-access-6hnpl\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.202138 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.202203 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.202421 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.202460 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4469e851-270f-47ee-a523-e3fb56d8201f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.308597 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.308649 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4469e851-270f-47ee-a523-e3fb56d8201f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.314306 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4469e851-270f-47ee-a523-e3fb56d8201f-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.317327 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.317923 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4469e851-270f-47ee-a523-e3fb56d8201f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318012 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318104 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318138 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-config\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318199 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318237 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hnpl\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-kube-api-access-6hnpl\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318282 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.318330 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.333675 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4469e851-270f-47ee-a523-e3fb56d8201f-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.334062 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.334241 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.334338 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.334479 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.335626 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-config\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.338349 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.340804 4829 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.340834 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf460093261147c62473ff98f8811184b7db6218b040c193e53546be3326fb82/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.350412 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hnpl\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-kube-api-access-6hnpl\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.351845 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.407981 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.562807 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.624960 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz82h\" (UniqueName: \"kubernetes.io/projected/9ae65aa7-5414-49a2-bd11-dd2209873f7e-kube-api-access-hz82h\") pod \"9ae65aa7-5414-49a2-bd11-dd2209873f7e\" (UID: \"9ae65aa7-5414-49a2-bd11-dd2209873f7e\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.629725 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ae65aa7-5414-49a2-bd11-dd2209873f7e-kube-api-access-hz82h" (OuterVolumeSpecName: "kube-api-access-hz82h") pod "9ae65aa7-5414-49a2-bd11-dd2209873f7e" (UID: "9ae65aa7-5414-49a2-bd11-dd2209873f7e"). InnerVolumeSpecName "kube-api-access-hz82h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.640752 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.694626 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726457 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw7v4\" (UniqueName: \"kubernetes.io/projected/b5063471-4946-4763-9338-1ad3fbb4b9ef-kube-api-access-dw7v4\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726562 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-ring-data-devices\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726623 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-scripts\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726661 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-dispersionconf\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726701 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-combined-ca-bundle\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726730 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5063471-4946-4763-9338-1ad3fbb4b9ef-etc-swift\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.726831 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-swiftconf\") pod \"b5063471-4946-4763-9338-1ad3fbb4b9ef\" (UID: \"b5063471-4946-4763-9338-1ad3fbb4b9ef\") " Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.727261 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz82h\" (UniqueName: \"kubernetes.io/projected/9ae65aa7-5414-49a2-bd11-dd2209873f7e-kube-api-access-hz82h\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.727282 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.727936 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5063471-4946-4763-9338-1ad3fbb4b9ef-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.731074 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5063471-4946-4763-9338-1ad3fbb4b9ef-kube-api-access-dw7v4" (OuterVolumeSpecName: "kube-api-access-dw7v4") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "kube-api-access-dw7v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.732417 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.753081 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-scripts" (OuterVolumeSpecName: "scripts") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.757038 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.759401 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "b5063471-4946-4763-9338-1ad3fbb4b9ef" (UID: "b5063471-4946-4763-9338-1ad3fbb4b9ef"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.828709 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw7v4\" (UniqueName: \"kubernetes.io/projected/b5063471-4946-4763-9338-1ad3fbb4b9ef-kube-api-access-dw7v4\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.829788 4829 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.829820 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b5063471-4946-4763-9338-1ad3fbb4b9ef-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.829835 4829 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.829851 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.829864 4829 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/b5063471-4946-4763-9338-1ad3fbb4b9ef-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.829878 4829 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/b5063471-4946-4763-9338-1ad3fbb4b9ef-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.973027 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4939-account-create-2wx9l" event={"ID":"9ae65aa7-5414-49a2-bd11-dd2209873f7e","Type":"ContainerDied","Data":"308b8c2ee2df531bec333613c7b7e09975d723127bc469ba469e10f6d9be9998"} Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.973056 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4939-account-create-2wx9l" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.973069 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="308b8c2ee2df531bec333613c7b7e09975d723127bc469ba469e10f6d9be9998" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.974334 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-m6p2d" event={"ID":"b5063471-4946-4763-9338-1ad3fbb4b9ef","Type":"ContainerDied","Data":"d608ca387bb1d5a254ad8d7a064225bcb5d407e9599f44a5eaf0d74e4028f6f7"} Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.974350 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d608ca387bb1d5a254ad8d7a064225bcb5d407e9599f44a5eaf0d74e4028f6f7" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.974388 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-m6p2d" Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.977285 4829 generic.go:334] "Generic (PLEG): container finished" podID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerID="3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7" exitCode=0 Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.977397 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2","Type":"ContainerDied","Data":"3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7"} Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.979895 4829 generic.go:334] "Generic (PLEG): container finished" podID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerID="fad8f75f06da64a1030f412dfcd4668bdcec0ea2d6b25739b69ddbef0345d357" exitCode=0 Oct 02 07:33:38 crc kubenswrapper[4829]: I1002 07:33:38.979932 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bf3dfd-bae4-4af6-a338-fad9857bfd29","Type":"ContainerDied","Data":"fad8f75f06da64a1030f412dfcd4668bdcec0ea2d6b25739b69ddbef0345d357"} Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.130470 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 07:33:39 crc kubenswrapper[4829]: W1002 07:33:39.135115 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4469e851_270f_47ee_a523_e3fb56d8201f.slice/crio-437d10f7d98185338bb1b1d5571e80974322d31b853f19d9456c319c916194fa WatchSource:0}: Error finding container 437d10f7d98185338bb1b1d5571e80974322d31b853f19d9456c319c916194fa: Status 404 returned error can't find the container with id 437d10f7d98185338bb1b1d5571e80974322d31b853f19d9456c319c916194fa Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.475545 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f93d4a-ce37-43ad-85d0-4be548d45c3b" path="/var/lib/kubelet/pods/61f93d4a-ce37-43ad-85d0-4be548d45c3b/volumes" Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.993647 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerStarted","Data":"437d10f7d98185338bb1b1d5571e80974322d31b853f19d9456c319c916194fa"} Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.996038 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2","Type":"ContainerStarted","Data":"270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e"} Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.996259 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.998685 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bf3dfd-bae4-4af6-a338-fad9857bfd29","Type":"ContainerStarted","Data":"47a462567327e8b2e3336281a0ec373ccb15e3712fdae76fb61751f95c28d012"} Oct 02 07:33:39 crc kubenswrapper[4829]: I1002 07:33:39.998935 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.030840 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.591531647 podStartE2EDuration="1m1.03082291s" podCreationTimestamp="2025-10-02 07:32:39 +0000 UTC" firstStartedPulling="2025-10-02 07:32:53.147368995 +0000 UTC m=+964.487017410" lastFinishedPulling="2025-10-02 07:33:03.586660268 +0000 UTC m=+974.926308673" observedRunningTime="2025-10-02 07:33:40.029959542 +0000 UTC m=+1011.369607947" watchObservedRunningTime="2025-10-02 07:33:40.03082291 +0000 UTC m=+1011.370471315" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.066090 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.145243289 podStartE2EDuration="1m1.06607391s" podCreationTimestamp="2025-10-02 07:32:39 +0000 UTC" firstStartedPulling="2025-10-02 07:32:52.920773962 +0000 UTC m=+964.260422367" lastFinishedPulling="2025-10-02 07:33:02.841604583 +0000 UTC m=+974.181252988" observedRunningTime="2025-10-02 07:33:40.061548163 +0000 UTC m=+1011.401196568" watchObservedRunningTime="2025-10-02 07:33:40.06607391 +0000 UTC m=+1011.405722315" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.234115 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-b5txm"] Oct 02 07:33:40 crc kubenswrapper[4829]: E1002 07:33:40.234830 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ae65aa7-5414-49a2-bd11-dd2209873f7e" containerName="mariadb-account-create" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.234852 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ae65aa7-5414-49a2-bd11-dd2209873f7e" containerName="mariadb-account-create" Oct 02 07:33:40 crc kubenswrapper[4829]: E1002 07:33:40.234888 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5063471-4946-4763-9338-1ad3fbb4b9ef" containerName="swift-ring-rebalance" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.234896 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5063471-4946-4763-9338-1ad3fbb4b9ef" containerName="swift-ring-rebalance" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.235124 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5063471-4946-4763-9338-1ad3fbb4b9ef" containerName="swift-ring-rebalance" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.235153 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ae65aa7-5414-49a2-bd11-dd2209873f7e" containerName="mariadb-account-create" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.235879 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.238033 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gtlmv" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.238307 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.242913 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-b5txm"] Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.251650 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-combined-ca-bundle\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.251737 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-db-sync-config-data\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.251761 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-config-data\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.251792 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwjnz\" (UniqueName: \"kubernetes.io/projected/97b06d08-abdf-4107-9561-9b81768e183c-kube-api-access-vwjnz\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.353601 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-combined-ca-bundle\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.353684 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-db-sync-config-data\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.353713 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-config-data\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.353745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwjnz\" (UniqueName: \"kubernetes.io/projected/97b06d08-abdf-4107-9561-9b81768e183c-kube-api-access-vwjnz\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.357757 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-db-sync-config-data\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.357884 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-combined-ca-bundle\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.359794 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-config-data\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.374721 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwjnz\" (UniqueName: \"kubernetes.io/projected/97b06d08-abdf-4107-9561-9b81768e183c-kube-api-access-vwjnz\") pod \"glance-db-sync-b5txm\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.555018 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b5txm" Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.618814 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-njw55" podUID="91ff1e5f-73e2-4237-b201-84065a586553" containerName="ovn-controller" probeResult="failure" output=< Oct 02 07:33:40 crc kubenswrapper[4829]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 07:33:40 crc kubenswrapper[4829]: > Oct 02 07:33:40 crc kubenswrapper[4829]: I1002 07:33:40.667259 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.353066 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-t8kr7" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.420270 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-b5txm"] Oct 02 07:33:41 crc kubenswrapper[4829]: W1002 07:33:41.422787 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97b06d08_abdf_4107_9561_9b81768e183c.slice/crio-6b8411efbe15d97335a6c2ade94d2b0c01ff730ce21f1b476ec3ee18f8b315a8 WatchSource:0}: Error finding container 6b8411efbe15d97335a6c2ade94d2b0c01ff730ce21f1b476ec3ee18f8b315a8: Status 404 returned error can't find the container with id 6b8411efbe15d97335a6c2ade94d2b0c01ff730ce21f1b476ec3ee18f8b315a8 Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.578123 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-njw55-config-n8vc5"] Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.585554 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.588902 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-njw55-config-n8vc5"] Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.593543 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.775937 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run-ovn\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.776004 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-log-ovn\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.776082 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlfvx\" (UniqueName: \"kubernetes.io/projected/f09576cb-229f-4756-bb06-37ecb00eaa72-kube-api-access-vlfvx\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.776101 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.776118 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-additional-scripts\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.776152 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-scripts\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.877269 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-scripts\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.877326 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run-ovn\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.877387 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-log-ovn\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.877468 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlfvx\" (UniqueName: \"kubernetes.io/projected/f09576cb-229f-4756-bb06-37ecb00eaa72-kube-api-access-vlfvx\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.877490 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.877506 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-additional-scripts\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.878262 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-additional-scripts\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.880211 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-scripts\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.880476 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run-ovn\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.880540 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-log-ovn\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.880811 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.900657 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlfvx\" (UniqueName: \"kubernetes.io/projected/f09576cb-229f-4756-bb06-37ecb00eaa72-kube-api-access-vlfvx\") pod \"ovn-controller-njw55-config-n8vc5\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:41 crc kubenswrapper[4829]: I1002 07:33:41.910918 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:42 crc kubenswrapper[4829]: I1002 07:33:42.027309 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b5txm" event={"ID":"97b06d08-abdf-4107-9561-9b81768e183c","Type":"ContainerStarted","Data":"6b8411efbe15d97335a6c2ade94d2b0c01ff730ce21f1b476ec3ee18f8b315a8"} Oct 02 07:33:42 crc kubenswrapper[4829]: I1002 07:33:42.381456 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-njw55-config-n8vc5"] Oct 02 07:33:42 crc kubenswrapper[4829]: W1002 07:33:42.405467 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf09576cb_229f_4756_bb06_37ecb00eaa72.slice/crio-8076b5e688485ae053256a486edffe930172c85b28acb4aa6918ad478f503cbc WatchSource:0}: Error finding container 8076b5e688485ae053256a486edffe930172c85b28acb4aa6918ad478f503cbc: Status 404 returned error can't find the container with id 8076b5e688485ae053256a486edffe930172c85b28acb4aa6918ad478f503cbc Oct 02 07:33:43 crc kubenswrapper[4829]: I1002 07:33:43.049503 4829 generic.go:334] "Generic (PLEG): container finished" podID="f09576cb-229f-4756-bb06-37ecb00eaa72" containerID="91bbbebef21bd98ba4714faafc6d34cf2ceca44eff5e5ec17afa2d02308bef43" exitCode=0 Oct 02 07:33:43 crc kubenswrapper[4829]: I1002 07:33:43.049555 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-njw55-config-n8vc5" event={"ID":"f09576cb-229f-4756-bb06-37ecb00eaa72","Type":"ContainerDied","Data":"91bbbebef21bd98ba4714faafc6d34cf2ceca44eff5e5ec17afa2d02308bef43"} Oct 02 07:33:43 crc kubenswrapper[4829]: I1002 07:33:43.050193 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-njw55-config-n8vc5" event={"ID":"f09576cb-229f-4756-bb06-37ecb00eaa72","Type":"ContainerStarted","Data":"8076b5e688485ae053256a486edffe930172c85b28acb4aa6918ad478f503cbc"} Oct 02 07:33:43 crc kubenswrapper[4829]: I1002 07:33:43.053950 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerStarted","Data":"50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6"} Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.392324 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6628-account-create-dwgd4"] Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.392726 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:44 crc kubenswrapper[4829]: E1002 07:33:44.393791 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09576cb-229f-4756-bb06-37ecb00eaa72" containerName="ovn-config" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.393815 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09576cb-229f-4756-bb06-37ecb00eaa72" containerName="ovn-config" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.394026 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09576cb-229f-4756-bb06-37ecb00eaa72" containerName="ovn-config" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.394775 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.397512 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.402618 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6628-account-create-dwgd4"] Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527506 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlfvx\" (UniqueName: \"kubernetes.io/projected/f09576cb-229f-4756-bb06-37ecb00eaa72-kube-api-access-vlfvx\") pod \"f09576cb-229f-4756-bb06-37ecb00eaa72\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527590 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-log-ovn\") pod \"f09576cb-229f-4756-bb06-37ecb00eaa72\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527682 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-additional-scripts\") pod \"f09576cb-229f-4756-bb06-37ecb00eaa72\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527736 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f09576cb-229f-4756-bb06-37ecb00eaa72" (UID: "f09576cb-229f-4756-bb06-37ecb00eaa72"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527811 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run\") pod \"f09576cb-229f-4756-bb06-37ecb00eaa72\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527830 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run" (OuterVolumeSpecName: "var-run") pod "f09576cb-229f-4756-bb06-37ecb00eaa72" (UID: "f09576cb-229f-4756-bb06-37ecb00eaa72"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527855 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-scripts\") pod \"f09576cb-229f-4756-bb06-37ecb00eaa72\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.527936 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run-ovn\") pod \"f09576cb-229f-4756-bb06-37ecb00eaa72\" (UID: \"f09576cb-229f-4756-bb06-37ecb00eaa72\") " Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528254 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f09576cb-229f-4756-bb06-37ecb00eaa72" (UID: "f09576cb-229f-4756-bb06-37ecb00eaa72"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528299 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f09576cb-229f-4756-bb06-37ecb00eaa72" (UID: "f09576cb-229f-4756-bb06-37ecb00eaa72"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528372 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkgwt\" (UniqueName: \"kubernetes.io/projected/e3fba779-519a-4959-860b-e6daf434d2ee-kube-api-access-dkgwt\") pod \"keystone-6628-account-create-dwgd4\" (UID: \"e3fba779-519a-4959-860b-e6daf434d2ee\") " pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528457 4829 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528472 4829 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528482 4829 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.528490 4829 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f09576cb-229f-4756-bb06-37ecb00eaa72-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.529559 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-scripts" (OuterVolumeSpecName: "scripts") pod "f09576cb-229f-4756-bb06-37ecb00eaa72" (UID: "f09576cb-229f-4756-bb06-37ecb00eaa72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.532851 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09576cb-229f-4756-bb06-37ecb00eaa72-kube-api-access-vlfvx" (OuterVolumeSpecName: "kube-api-access-vlfvx") pod "f09576cb-229f-4756-bb06-37ecb00eaa72" (UID: "f09576cb-229f-4756-bb06-37ecb00eaa72"). InnerVolumeSpecName "kube-api-access-vlfvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.630088 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkgwt\" (UniqueName: \"kubernetes.io/projected/e3fba779-519a-4959-860b-e6daf434d2ee-kube-api-access-dkgwt\") pod \"keystone-6628-account-create-dwgd4\" (UID: \"e3fba779-519a-4959-860b-e6daf434d2ee\") " pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.630259 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f09576cb-229f-4756-bb06-37ecb00eaa72-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.630274 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlfvx\" (UniqueName: \"kubernetes.io/projected/f09576cb-229f-4756-bb06-37ecb00eaa72-kube-api-access-vlfvx\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.660154 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkgwt\" (UniqueName: \"kubernetes.io/projected/e3fba779-519a-4959-860b-e6daf434d2ee-kube-api-access-dkgwt\") pod \"keystone-6628-account-create-dwgd4\" (UID: \"e3fba779-519a-4959-860b-e6daf434d2ee\") " pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:44 crc kubenswrapper[4829]: I1002 07:33:44.708425 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.072839 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-njw55-config-n8vc5" event={"ID":"f09576cb-229f-4756-bb06-37ecb00eaa72","Type":"ContainerDied","Data":"8076b5e688485ae053256a486edffe930172c85b28acb4aa6918ad478f503cbc"} Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.073211 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8076b5e688485ae053256a486edffe930172c85b28acb4aa6918ad478f503cbc" Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.072945 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-njw55-config-n8vc5" Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.161205 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6628-account-create-dwgd4"] Oct 02 07:33:45 crc kubenswrapper[4829]: W1002 07:33:45.161992 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3fba779_519a_4959_860b_e6daf434d2ee.slice/crio-a592add5fd12792e3e4d700bb20243f49471c7ad39ea6b2f5183b855e270ad5d WatchSource:0}: Error finding container a592add5fd12792e3e4d700bb20243f49471c7ad39ea6b2f5183b855e270ad5d: Status 404 returned error can't find the container with id a592add5fd12792e3e4d700bb20243f49471c7ad39ea6b2f5183b855e270ad5d Oct 02 07:33:45 crc kubenswrapper[4829]: E1002 07:33:45.272343 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf09576cb_229f_4756_bb06_37ecb00eaa72.slice\": RecentStats: unable to find data in memory cache]" Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.493754 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-njw55-config-n8vc5"] Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.501789 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-njw55-config-n8vc5"] Oct 02 07:33:45 crc kubenswrapper[4829]: I1002 07:33:45.572615 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-njw55" Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.113975 4829 generic.go:334] "Generic (PLEG): container finished" podID="e3fba779-519a-4959-860b-e6daf434d2ee" containerID="7f348771e01df8391a5d48147093da9cd70eda5bd6b1bbb8a86a7bd21cdb06ba" exitCode=0 Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.114063 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6628-account-create-dwgd4" event={"ID":"e3fba779-519a-4959-860b-e6daf434d2ee","Type":"ContainerDied","Data":"7f348771e01df8391a5d48147093da9cd70eda5bd6b1bbb8a86a7bd21cdb06ba"} Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.114149 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6628-account-create-dwgd4" event={"ID":"e3fba779-519a-4959-860b-e6daf434d2ee","Type":"ContainerStarted","Data":"a592add5fd12792e3e4d700bb20243f49471c7ad39ea6b2f5183b855e270ad5d"} Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.649804 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db66-account-create-ldcv7"] Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.652861 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.654796 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.661017 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db66-account-create-ldcv7"] Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.764856 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lvk4\" (UniqueName: \"kubernetes.io/projected/6e9830ff-3291-43f6-89cf-3911036a3350-kube-api-access-5lvk4\") pod \"watcher-db66-account-create-ldcv7\" (UID: \"6e9830ff-3291-43f6-89cf-3911036a3350\") " pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.866473 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lvk4\" (UniqueName: \"kubernetes.io/projected/6e9830ff-3291-43f6-89cf-3911036a3350-kube-api-access-5lvk4\") pod \"watcher-db66-account-create-ldcv7\" (UID: \"6e9830ff-3291-43f6-89cf-3911036a3350\") " pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:46 crc kubenswrapper[4829]: I1002 07:33:46.885870 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lvk4\" (UniqueName: \"kubernetes.io/projected/6e9830ff-3291-43f6-89cf-3911036a3350-kube-api-access-5lvk4\") pod \"watcher-db66-account-create-ldcv7\" (UID: \"6e9830ff-3291-43f6-89cf-3911036a3350\") " pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:47 crc kubenswrapper[4829]: I1002 07:33:47.001035 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:47 crc kubenswrapper[4829]: I1002 07:33:47.475829 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f09576cb-229f-4756-bb06-37ecb00eaa72" path="/var/lib/kubelet/pods/f09576cb-229f-4756-bb06-37ecb00eaa72/volumes" Oct 02 07:33:49 crc kubenswrapper[4829]: I1002 07:33:49.146931 4829 generic.go:334] "Generic (PLEG): container finished" podID="4469e851-270f-47ee-a523-e3fb56d8201f" containerID="50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6" exitCode=0 Oct 02 07:33:49 crc kubenswrapper[4829]: I1002 07:33:49.147039 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerDied","Data":"50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6"} Oct 02 07:33:49 crc kubenswrapper[4829]: I1002 07:33:49.617822 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:49 crc kubenswrapper[4829]: I1002 07:33:49.628562 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9da44bca-624e-49ca-8fd2-e0b0974c9ae5-etc-swift\") pod \"swift-storage-0\" (UID: \"9da44bca-624e-49ca-8fd2-e0b0974c9ae5\") " pod="openstack/swift-storage-0" Oct 02 07:33:49 crc kubenswrapper[4829]: I1002 07:33:49.861584 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 07:33:51 crc kubenswrapper[4829]: I1002 07:33:51.078387 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:33:51 crc kubenswrapper[4829]: I1002 07:33:51.406402 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 07:33:52 crc kubenswrapper[4829]: I1002 07:33:52.826826 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-zrcg4"] Oct 02 07:33:52 crc kubenswrapper[4829]: I1002 07:33:52.828124 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:52 crc kubenswrapper[4829]: I1002 07:33:52.834894 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zrcg4"] Oct 02 07:33:52 crc kubenswrapper[4829]: I1002 07:33:52.874271 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nf46\" (UniqueName: \"kubernetes.io/projected/61eaa9fd-41d9-4341-8f74-acef1d2782d6-kube-api-access-6nf46\") pod \"barbican-db-create-zrcg4\" (UID: \"61eaa9fd-41d9-4341-8f74-acef1d2782d6\") " pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:52 crc kubenswrapper[4829]: I1002 07:33:52.975445 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nf46\" (UniqueName: \"kubernetes.io/projected/61eaa9fd-41d9-4341-8f74-acef1d2782d6-kube-api-access-6nf46\") pod \"barbican-db-create-zrcg4\" (UID: \"61eaa9fd-41d9-4341-8f74-acef1d2782d6\") " pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:52 crc kubenswrapper[4829]: I1002 07:33:52.996848 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nf46\" (UniqueName: \"kubernetes.io/projected/61eaa9fd-41d9-4341-8f74-acef1d2782d6-kube-api-access-6nf46\") pod \"barbican-db-create-zrcg4\" (UID: \"61eaa9fd-41d9-4341-8f74-acef1d2782d6\") " pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.040142 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-s25rw"] Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.041142 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.048076 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-s25rw"] Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.129610 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-z22p2"] Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.132118 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.142718 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.147286 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z22p2"] Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.178868 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npxhw\" (UniqueName: \"kubernetes.io/projected/c61c72f6-3856-4475-9d67-eaa307c9d23e-kube-api-access-npxhw\") pod \"neutron-db-create-z22p2\" (UID: \"c61c72f6-3856-4475-9d67-eaa307c9d23e\") " pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.179005 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt287\" (UniqueName: \"kubernetes.io/projected/9df948cb-5fcf-4d8c-868d-4d80cbe4c172-kube-api-access-lt287\") pod \"cinder-db-create-s25rw\" (UID: \"9df948cb-5fcf-4d8c-868d-4d80cbe4c172\") " pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.280612 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt287\" (UniqueName: \"kubernetes.io/projected/9df948cb-5fcf-4d8c-868d-4d80cbe4c172-kube-api-access-lt287\") pod \"cinder-db-create-s25rw\" (UID: \"9df948cb-5fcf-4d8c-868d-4d80cbe4c172\") " pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.280695 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npxhw\" (UniqueName: \"kubernetes.io/projected/c61c72f6-3856-4475-9d67-eaa307c9d23e-kube-api-access-npxhw\") pod \"neutron-db-create-z22p2\" (UID: \"c61c72f6-3856-4475-9d67-eaa307c9d23e\") " pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.307217 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npxhw\" (UniqueName: \"kubernetes.io/projected/c61c72f6-3856-4475-9d67-eaa307c9d23e-kube-api-access-npxhw\") pod \"neutron-db-create-z22p2\" (UID: \"c61c72f6-3856-4475-9d67-eaa307c9d23e\") " pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.308703 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt287\" (UniqueName: \"kubernetes.io/projected/9df948cb-5fcf-4d8c-868d-4d80cbe4c172-kube-api-access-lt287\") pod \"cinder-db-create-s25rw\" (UID: \"9df948cb-5fcf-4d8c-868d-4d80cbe4c172\") " pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.355572 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:53 crc kubenswrapper[4829]: I1002 07:33:53.447529 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:54 crc kubenswrapper[4829]: I1002 07:33:54.383953 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:54 crc kubenswrapper[4829]: I1002 07:33:54.506472 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkgwt\" (UniqueName: \"kubernetes.io/projected/e3fba779-519a-4959-860b-e6daf434d2ee-kube-api-access-dkgwt\") pod \"e3fba779-519a-4959-860b-e6daf434d2ee\" (UID: \"e3fba779-519a-4959-860b-e6daf434d2ee\") " Oct 02 07:33:54 crc kubenswrapper[4829]: I1002 07:33:54.513180 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3fba779-519a-4959-860b-e6daf434d2ee-kube-api-access-dkgwt" (OuterVolumeSpecName: "kube-api-access-dkgwt") pod "e3fba779-519a-4959-860b-e6daf434d2ee" (UID: "e3fba779-519a-4959-860b-e6daf434d2ee"). InnerVolumeSpecName "kube-api-access-dkgwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:54 crc kubenswrapper[4829]: I1002 07:33:54.608038 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkgwt\" (UniqueName: \"kubernetes.io/projected/e3fba779-519a-4959-860b-e6daf434d2ee-kube-api-access-dkgwt\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:54 crc kubenswrapper[4829]: I1002 07:33:54.928166 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zrcg4"] Oct 02 07:33:54 crc kubenswrapper[4829]: W1002 07:33:54.947590 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod61eaa9fd_41d9_4341_8f74_acef1d2782d6.slice/crio-a7f16359f922d390c50cca776f09e4707ca51ef1c4004248686fffc1ef623914 WatchSource:0}: Error finding container a7f16359f922d390c50cca776f09e4707ca51ef1c4004248686fffc1ef623914: Status 404 returned error can't find the container with id a7f16359f922d390c50cca776f09e4707ca51ef1c4004248686fffc1ef623914 Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.056531 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-z22p2"] Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.198042 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 07:33:55 crc kubenswrapper[4829]: W1002 07:33:55.225248 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9da44bca_624e_49ca_8fd2_e0b0974c9ae5.slice/crio-3ca0bd52f279db2b0c056f6a4ea7d4dec48a315daf7dd733602882e140889655 WatchSource:0}: Error finding container 3ca0bd52f279db2b0c056f6a4ea7d4dec48a315daf7dd733602882e140889655: Status 404 returned error can't find the container with id 3ca0bd52f279db2b0c056f6a4ea7d4dec48a315daf7dd733602882e140889655 Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.234593 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-s25rw"] Oct 02 07:33:55 crc kubenswrapper[4829]: W1002 07:33:55.237838 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e9830ff_3291_43f6_89cf_3911036a3350.slice/crio-142ca549f4d24e1657884eeac35935af41783f185101af9812d447a83bbf48b2 WatchSource:0}: Error finding container 142ca549f4d24e1657884eeac35935af41783f185101af9812d447a83bbf48b2: Status 404 returned error can't find the container with id 142ca549f4d24e1657884eeac35935af41783f185101af9812d447a83bbf48b2 Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.241968 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.245204 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerStarted","Data":"b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407"} Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.247032 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db66-account-create-ldcv7"] Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.247353 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z22p2" event={"ID":"c61c72f6-3856-4475-9d67-eaa307c9d23e","Type":"ContainerStarted","Data":"622039d96fe3d1c711241978b72fe3f1311062ec0f68e73b11ad1e010d80fbea"} Oct 02 07:33:55 crc kubenswrapper[4829]: W1002 07:33:55.248903 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9df948cb_5fcf_4d8c_868d_4d80cbe4c172.slice/crio-149b582ecd9755a5b3aad3f2f2f2d16ff7ec54a4a1bf86705daad1a398ab5ee6 WatchSource:0}: Error finding container 149b582ecd9755a5b3aad3f2f2f2d16ff7ec54a4a1bf86705daad1a398ab5ee6: Status 404 returned error can't find the container with id 149b582ecd9755a5b3aad3f2f2f2d16ff7ec54a4a1bf86705daad1a398ab5ee6 Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.251337 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"3ca0bd52f279db2b0c056f6a4ea7d4dec48a315daf7dd733602882e140889655"} Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.253689 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6628-account-create-dwgd4" event={"ID":"e3fba779-519a-4959-860b-e6daf434d2ee","Type":"ContainerDied","Data":"a592add5fd12792e3e4d700bb20243f49471c7ad39ea6b2f5183b855e270ad5d"} Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.253718 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a592add5fd12792e3e4d700bb20243f49471c7ad39ea6b2f5183b855e270ad5d" Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.253767 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6628-account-create-dwgd4" Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.267750 4829 generic.go:334] "Generic (PLEG): container finished" podID="61eaa9fd-41d9-4341-8f74-acef1d2782d6" containerID="17ede04b6f9b8448f6085cd42c78d9aedd5c585a76ef015f6cf4aecbed8a0c76" exitCode=0 Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.267799 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zrcg4" event={"ID":"61eaa9fd-41d9-4341-8f74-acef1d2782d6","Type":"ContainerDied","Data":"17ede04b6f9b8448f6085cd42c78d9aedd5c585a76ef015f6cf4aecbed8a0c76"} Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.267825 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zrcg4" event={"ID":"61eaa9fd-41d9-4341-8f74-acef1d2782d6","Type":"ContainerStarted","Data":"a7f16359f922d390c50cca776f09e4707ca51ef1c4004248686fffc1ef623914"} Oct 02 07:33:55 crc kubenswrapper[4829]: I1002 07:33:55.294310 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-z22p2" podStartSLOduration=2.294289062 podStartE2EDuration="2.294289062s" podCreationTimestamp="2025-10-02 07:33:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:55.265670397 +0000 UTC m=+1026.605318802" watchObservedRunningTime="2025-10-02 07:33:55.294289062 +0000 UTC m=+1026.633937467" Oct 02 07:33:55 crc kubenswrapper[4829]: E1002 07:33:55.524274 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc61c72f6_3856_4475_9d67_eaa307c9d23e.slice/crio-2ee0d197c56046778ed4d28f0d37be24fddd763c1bbfb976c55c42dc784374ef.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.300995 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b5txm" event={"ID":"97b06d08-abdf-4107-9561-9b81768e183c","Type":"ContainerStarted","Data":"bf6db1a9b6aedd6448837ea1085c40b0a89f07480ad31e3dc8203c881da5fe5e"} Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.316996 4829 generic.go:334] "Generic (PLEG): container finished" podID="6e9830ff-3291-43f6-89cf-3911036a3350" containerID="484e8753483fb11a7420eb11926e3496f314702908fcb1ee6f97d0fe9c517ce9" exitCode=0 Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.317080 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db66-account-create-ldcv7" event={"ID":"6e9830ff-3291-43f6-89cf-3911036a3350","Type":"ContainerDied","Data":"484e8753483fb11a7420eb11926e3496f314702908fcb1ee6f97d0fe9c517ce9"} Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.317105 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db66-account-create-ldcv7" event={"ID":"6e9830ff-3291-43f6-89cf-3911036a3350","Type":"ContainerStarted","Data":"142ca549f4d24e1657884eeac35935af41783f185101af9812d447a83bbf48b2"} Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.325425 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-b5txm" podStartSLOduration=3.26968859 podStartE2EDuration="16.325409751s" podCreationTimestamp="2025-10-02 07:33:40 +0000 UTC" firstStartedPulling="2025-10-02 07:33:41.424871767 +0000 UTC m=+1012.764520172" lastFinishedPulling="2025-10-02 07:33:54.480592928 +0000 UTC m=+1025.820241333" observedRunningTime="2025-10-02 07:33:56.325008859 +0000 UTC m=+1027.664657274" watchObservedRunningTime="2025-10-02 07:33:56.325409751 +0000 UTC m=+1027.665058156" Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.329640 4829 generic.go:334] "Generic (PLEG): container finished" podID="c61c72f6-3856-4475-9d67-eaa307c9d23e" containerID="2ee0d197c56046778ed4d28f0d37be24fddd763c1bbfb976c55c42dc784374ef" exitCode=0 Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.329694 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z22p2" event={"ID":"c61c72f6-3856-4475-9d67-eaa307c9d23e","Type":"ContainerDied","Data":"2ee0d197c56046778ed4d28f0d37be24fddd763c1bbfb976c55c42dc784374ef"} Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.362367 4829 generic.go:334] "Generic (PLEG): container finished" podID="9df948cb-5fcf-4d8c-868d-4d80cbe4c172" containerID="6a1b99e31061ef8934d07fedcbaf6f21a33eef475ba6b90c1337e3e531d8d40f" exitCode=0 Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.362605 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s25rw" event={"ID":"9df948cb-5fcf-4d8c-868d-4d80cbe4c172","Type":"ContainerDied","Data":"6a1b99e31061ef8934d07fedcbaf6f21a33eef475ba6b90c1337e3e531d8d40f"} Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.362628 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s25rw" event={"ID":"9df948cb-5fcf-4d8c-868d-4d80cbe4c172","Type":"ContainerStarted","Data":"149b582ecd9755a5b3aad3f2f2f2d16ff7ec54a4a1bf86705daad1a398ab5ee6"} Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.815283 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:56 crc kubenswrapper[4829]: I1002 07:33:56.973580 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nf46\" (UniqueName: \"kubernetes.io/projected/61eaa9fd-41d9-4341-8f74-acef1d2782d6-kube-api-access-6nf46\") pod \"61eaa9fd-41d9-4341-8f74-acef1d2782d6\" (UID: \"61eaa9fd-41d9-4341-8f74-acef1d2782d6\") " Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.092389 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61eaa9fd-41d9-4341-8f74-acef1d2782d6-kube-api-access-6nf46" (OuterVolumeSpecName: "kube-api-access-6nf46") pod "61eaa9fd-41d9-4341-8f74-acef1d2782d6" (UID: "61eaa9fd-41d9-4341-8f74-acef1d2782d6"). InnerVolumeSpecName "kube-api-access-6nf46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.177394 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nf46\" (UniqueName: \"kubernetes.io/projected/61eaa9fd-41d9-4341-8f74-acef1d2782d6-kube-api-access-6nf46\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.374299 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zrcg4" event={"ID":"61eaa9fd-41d9-4341-8f74-acef1d2782d6","Type":"ContainerDied","Data":"a7f16359f922d390c50cca776f09e4707ca51ef1c4004248686fffc1ef623914"} Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.374365 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7f16359f922d390c50cca776f09e4707ca51ef1c4004248686fffc1ef623914" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.374474 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zrcg4" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.829511 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.966963 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.975889 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:57 crc kubenswrapper[4829]: I1002 07:33:57.992492 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npxhw\" (UniqueName: \"kubernetes.io/projected/c61c72f6-3856-4475-9d67-eaa307c9d23e-kube-api-access-npxhw\") pod \"c61c72f6-3856-4475-9d67-eaa307c9d23e\" (UID: \"c61c72f6-3856-4475-9d67-eaa307c9d23e\") " Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.064666 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c61c72f6-3856-4475-9d67-eaa307c9d23e-kube-api-access-npxhw" (OuterVolumeSpecName: "kube-api-access-npxhw") pod "c61c72f6-3856-4475-9d67-eaa307c9d23e" (UID: "c61c72f6-3856-4475-9d67-eaa307c9d23e"). InnerVolumeSpecName "kube-api-access-npxhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.094644 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt287\" (UniqueName: \"kubernetes.io/projected/9df948cb-5fcf-4d8c-868d-4d80cbe4c172-kube-api-access-lt287\") pod \"9df948cb-5fcf-4d8c-868d-4d80cbe4c172\" (UID: \"9df948cb-5fcf-4d8c-868d-4d80cbe4c172\") " Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.094899 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lvk4\" (UniqueName: \"kubernetes.io/projected/6e9830ff-3291-43f6-89cf-3911036a3350-kube-api-access-5lvk4\") pod \"6e9830ff-3291-43f6-89cf-3911036a3350\" (UID: \"6e9830ff-3291-43f6-89cf-3911036a3350\") " Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.095492 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npxhw\" (UniqueName: \"kubernetes.io/projected/c61c72f6-3856-4475-9d67-eaa307c9d23e-kube-api-access-npxhw\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.099326 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e9830ff-3291-43f6-89cf-3911036a3350-kube-api-access-5lvk4" (OuterVolumeSpecName: "kube-api-access-5lvk4") pod "6e9830ff-3291-43f6-89cf-3911036a3350" (UID: "6e9830ff-3291-43f6-89cf-3911036a3350"). InnerVolumeSpecName "kube-api-access-5lvk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.099592 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9df948cb-5fcf-4d8c-868d-4d80cbe4c172-kube-api-access-lt287" (OuterVolumeSpecName: "kube-api-access-lt287") pod "9df948cb-5fcf-4d8c-868d-4d80cbe4c172" (UID: "9df948cb-5fcf-4d8c-868d-4d80cbe4c172"). InnerVolumeSpecName "kube-api-access-lt287". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.197107 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt287\" (UniqueName: \"kubernetes.io/projected/9df948cb-5fcf-4d8c-868d-4d80cbe4c172-kube-api-access-lt287\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.197147 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lvk4\" (UniqueName: \"kubernetes.io/projected/6e9830ff-3291-43f6-89cf-3911036a3350-kube-api-access-5lvk4\") on node \"crc\" DevicePath \"\"" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.400632 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db66-account-create-ldcv7" event={"ID":"6e9830ff-3291-43f6-89cf-3911036a3350","Type":"ContainerDied","Data":"142ca549f4d24e1657884eeac35935af41783f185101af9812d447a83bbf48b2"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.400688 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="142ca549f4d24e1657884eeac35935af41783f185101af9812d447a83bbf48b2" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.400659 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db66-account-create-ldcv7" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.404068 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-z22p2" event={"ID":"c61c72f6-3856-4475-9d67-eaa307c9d23e","Type":"ContainerDied","Data":"622039d96fe3d1c711241978b72fe3f1311062ec0f68e73b11ad1e010d80fbea"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.404108 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622039d96fe3d1c711241978b72fe3f1311062ec0f68e73b11ad1e010d80fbea" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.404186 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-z22p2" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.406914 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-s25rw" event={"ID":"9df948cb-5fcf-4d8c-868d-4d80cbe4c172","Type":"ContainerDied","Data":"149b582ecd9755a5b3aad3f2f2f2d16ff7ec54a4a1bf86705daad1a398ab5ee6"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.406986 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="149b582ecd9755a5b3aad3f2f2f2d16ff7ec54a4a1bf86705daad1a398ab5ee6" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.407072 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-s25rw" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.410293 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"3e28c574d4cfe789b8444dee8479198388e6d9e0a89c9de279d9dd72e75af7be"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.410347 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"9fc9ec651519ed2386ba0548fe28cbc4162512a3d3625a1f67a332f41d2ffff0"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.410358 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"ee090c83a0e99c1b4ac01074e3477f2cb52cbeb42915ceee83d7975248efad55"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.413484 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerStarted","Data":"83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.413531 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerStarted","Data":"9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247"} Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.448980 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.448964031 podStartE2EDuration="20.448964031s" podCreationTimestamp="2025-10-02 07:33:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:33:58.443313948 +0000 UTC m=+1029.782962423" watchObservedRunningTime="2025-10-02 07:33:58.448964031 +0000 UTC m=+1029.788612436" Oct 02 07:33:58 crc kubenswrapper[4829]: I1002 07:33:58.695813 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.430513 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"46c8c619f05a8536295fc7f1a9ab44e93bc19762ee55514269220232aad36de4"} Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.980937 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-89hnk"] Oct 02 07:33:59 crc kubenswrapper[4829]: E1002 07:33:59.981283 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9df948cb-5fcf-4d8c-868d-4d80cbe4c172" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981296 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9df948cb-5fcf-4d8c-868d-4d80cbe4c172" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: E1002 07:33:59.981308 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9830ff-3291-43f6-89cf-3911036a3350" containerName="mariadb-account-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981313 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9830ff-3291-43f6-89cf-3911036a3350" containerName="mariadb-account-create" Oct 02 07:33:59 crc kubenswrapper[4829]: E1002 07:33:59.981327 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c61c72f6-3856-4475-9d67-eaa307c9d23e" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981334 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c61c72f6-3856-4475-9d67-eaa307c9d23e" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: E1002 07:33:59.981348 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61eaa9fd-41d9-4341-8f74-acef1d2782d6" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981355 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="61eaa9fd-41d9-4341-8f74-acef1d2782d6" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: E1002 07:33:59.981364 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3fba779-519a-4959-860b-e6daf434d2ee" containerName="mariadb-account-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981370 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3fba779-519a-4959-860b-e6daf434d2ee" containerName="mariadb-account-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981529 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9df948cb-5fcf-4d8c-868d-4d80cbe4c172" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981542 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="61eaa9fd-41d9-4341-8f74-acef1d2782d6" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981554 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3fba779-519a-4959-860b-e6daf434d2ee" containerName="mariadb-account-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981570 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9830ff-3291-43f6-89cf-3911036a3350" containerName="mariadb-account-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.981582 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c61c72f6-3856-4475-9d67-eaa307c9d23e" containerName="mariadb-database-create" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.982090 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-89hnk" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.984305 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.984652 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.984781 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.996276 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-89hnk"] Oct 02 07:33:59 crc kubenswrapper[4829]: I1002 07:33:59.997680 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxrff" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.032262 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkqkm\" (UniqueName: \"kubernetes.io/projected/98649236-1856-4a42-8171-9435997ecb47-kube-api-access-zkqkm\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.032454 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-config-data\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.032515 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-combined-ca-bundle\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.133805 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-config-data\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.133891 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-combined-ca-bundle\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.133928 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkqkm\" (UniqueName: \"kubernetes.io/projected/98649236-1856-4a42-8171-9435997ecb47-kube-api-access-zkqkm\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.138791 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-config-data\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.138829 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-combined-ca-bundle\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.154772 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkqkm\" (UniqueName: \"kubernetes.io/projected/98649236-1856-4a42-8171-9435997ecb47-kube-api-access-zkqkm\") pod \"keystone-db-sync-89hnk\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.298618 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:00 crc kubenswrapper[4829]: I1002 07:34:00.928039 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-89hnk"] Oct 02 07:34:00 crc kubenswrapper[4829]: W1002 07:34:00.937740 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98649236_1856_4a42_8171_9435997ecb47.slice/crio-b6570d3ee095dd877ce2a17f923412f5dd8a0c9441c3ffd26302870fcd20af31 WatchSource:0}: Error finding container b6570d3ee095dd877ce2a17f923412f5dd8a0c9441c3ffd26302870fcd20af31: Status 404 returned error can't find the container with id b6570d3ee095dd877ce2a17f923412f5dd8a0c9441c3ffd26302870fcd20af31 Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.513426 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"512cedeac1c3f786919009a072f3a613380355f5261c5c1262aee8a89509489b"} Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.513816 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"159dc0c948ead57b29f65e51596950a7322ba068bcf09ae678e641b21c9d39ee"} Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.513834 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"3f514e52be0e3aad834adf926880aa52aa209a6dc1bde5bf4c2d5f72bc947157"} Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.513846 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-89hnk" event={"ID":"98649236-1856-4a42-8171-9435997ecb47","Type":"ContainerStarted","Data":"b6570d3ee095dd877ce2a17f923412f5dd8a0c9441c3ffd26302870fcd20af31"} Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.844369 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-28kmg"] Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.845381 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.848267 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-jksjr" Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.848280 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.862944 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-28kmg"] Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.970690 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-combined-ca-bundle\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.970828 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-db-sync-config-data\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.970888 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-config-data\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:01 crc kubenswrapper[4829]: I1002 07:34:01.970928 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c42qk\" (UniqueName: \"kubernetes.io/projected/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-kube-api-access-c42qk\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.072343 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-config-data\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.072416 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c42qk\" (UniqueName: \"kubernetes.io/projected/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-kube-api-access-c42qk\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.072485 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-combined-ca-bundle\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.072578 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-db-sync-config-data\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.078859 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-db-sync-config-data\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.080717 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-config-data\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.088784 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-combined-ca-bundle\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.089400 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c42qk\" (UniqueName: \"kubernetes.io/projected/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-kube-api-access-c42qk\") pod \"watcher-db-sync-28kmg\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:02 crc kubenswrapper[4829]: I1002 07:34:02.162606 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.489973 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"fd13f65569d4050126c162de4d352a0dad050b2dd96299bfd0adbabf0b097519"} Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.688388 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-28kmg"] Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.820916 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-3056-account-create-gdp9h"] Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.822062 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.825468 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.839769 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3056-account-create-gdp9h"] Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.887657 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7r89\" (UniqueName: \"kubernetes.io/projected/ebc9e693-09fe-4577-a32b-cad2af43ab8a-kube-api-access-t7r89\") pod \"barbican-3056-account-create-gdp9h\" (UID: \"ebc9e693-09fe-4577-a32b-cad2af43ab8a\") " pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:02.989871 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7r89\" (UniqueName: \"kubernetes.io/projected/ebc9e693-09fe-4577-a32b-cad2af43ab8a-kube-api-access-t7r89\") pod \"barbican-3056-account-create-gdp9h\" (UID: \"ebc9e693-09fe-4577-a32b-cad2af43ab8a\") " pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.008754 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-676c-account-create-6wfbq"] Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.010538 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.012312 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7r89\" (UniqueName: \"kubernetes.io/projected/ebc9e693-09fe-4577-a32b-cad2af43ab8a-kube-api-access-t7r89\") pod \"barbican-3056-account-create-gdp9h\" (UID: \"ebc9e693-09fe-4577-a32b-cad2af43ab8a\") " pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.014264 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.018645 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-676c-account-create-6wfbq"] Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.092311 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szcvc\" (UniqueName: \"kubernetes.io/projected/0f1fa854-ffef-4cfe-a58f-7fc7f382af72-kube-api-access-szcvc\") pod \"neutron-676c-account-create-6wfbq\" (UID: \"0f1fa854-ffef-4cfe-a58f-7fc7f382af72\") " pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.188880 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.193877 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szcvc\" (UniqueName: \"kubernetes.io/projected/0f1fa854-ffef-4cfe-a58f-7fc7f382af72-kube-api-access-szcvc\") pod \"neutron-676c-account-create-6wfbq\" (UID: \"0f1fa854-ffef-4cfe-a58f-7fc7f382af72\") " pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.211403 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szcvc\" (UniqueName: \"kubernetes.io/projected/0f1fa854-ffef-4cfe-a58f-7fc7f382af72-kube-api-access-szcvc\") pod \"neutron-676c-account-create-6wfbq\" (UID: \"0f1fa854-ffef-4cfe-a58f-7fc7f382af72\") " pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.366800 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.514838 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-28kmg" event={"ID":"af15cc1d-88dc-44f9-b43c-9570dfc4fd65","Type":"ContainerStarted","Data":"7dc4e2252ac561c9d5f40fcc6ef8312500f0292e14108bf52f58ddd3cfa5d3cc"} Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.533121 4829 generic.go:334] "Generic (PLEG): container finished" podID="97b06d08-abdf-4107-9561-9b81768e183c" containerID="bf6db1a9b6aedd6448837ea1085c40b0a89f07480ad31e3dc8203c881da5fe5e" exitCode=0 Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.533163 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b5txm" event={"ID":"97b06d08-abdf-4107-9561-9b81768e183c","Type":"ContainerDied","Data":"bf6db1a9b6aedd6448837ea1085c40b0a89f07480ad31e3dc8203c881da5fe5e"} Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.810081 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-3056-account-create-gdp9h"] Oct 02 07:34:03 crc kubenswrapper[4829]: W1002 07:34:03.825791 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebc9e693_09fe_4577_a32b_cad2af43ab8a.slice/crio-2712aab3cd8be7b517d9e50645921c7bc1e041ab53b042500606df59afd75198 WatchSource:0}: Error finding container 2712aab3cd8be7b517d9e50645921c7bc1e041ab53b042500606df59afd75198: Status 404 returned error can't find the container with id 2712aab3cd8be7b517d9e50645921c7bc1e041ab53b042500606df59afd75198 Oct 02 07:34:03 crc kubenswrapper[4829]: I1002 07:34:03.979726 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-676c-account-create-6wfbq"] Oct 02 07:34:03 crc kubenswrapper[4829]: W1002 07:34:03.989446 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f1fa854_ffef_4cfe_a58f_7fc7f382af72.slice/crio-6ebe6f1994fb90f8a63ba2a0c0fa64f976ba9ae9e75606d3ff5d4b22c00fdaa6 WatchSource:0}: Error finding container 6ebe6f1994fb90f8a63ba2a0c0fa64f976ba9ae9e75606d3ff5d4b22c00fdaa6: Status 404 returned error can't find the container with id 6ebe6f1994fb90f8a63ba2a0c0fa64f976ba9ae9e75606d3ff5d4b22c00fdaa6 Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.541961 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-676c-account-create-6wfbq" event={"ID":"0f1fa854-ffef-4cfe-a58f-7fc7f382af72","Type":"ContainerStarted","Data":"51fae9a9d6dde21c2c6b451d316fee6b2f43669963e47a89ea91d44ec5510306"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.542003 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-676c-account-create-6wfbq" event={"ID":"0f1fa854-ffef-4cfe-a58f-7fc7f382af72","Type":"ContainerStarted","Data":"6ebe6f1994fb90f8a63ba2a0c0fa64f976ba9ae9e75606d3ff5d4b22c00fdaa6"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.543543 4829 generic.go:334] "Generic (PLEG): container finished" podID="ebc9e693-09fe-4577-a32b-cad2af43ab8a" containerID="32cad8b33fedf01de6a1920e2cf46d92655ab50de25f5d22435a741bb4077496" exitCode=0 Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.543580 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3056-account-create-gdp9h" event={"ID":"ebc9e693-09fe-4577-a32b-cad2af43ab8a","Type":"ContainerDied","Data":"32cad8b33fedf01de6a1920e2cf46d92655ab50de25f5d22435a741bb4077496"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.543593 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3056-account-create-gdp9h" event={"ID":"ebc9e693-09fe-4577-a32b-cad2af43ab8a","Type":"ContainerStarted","Data":"2712aab3cd8be7b517d9e50645921c7bc1e041ab53b042500606df59afd75198"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.550727 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"c5f470996be5c71eec1d0d5007f419079f6973ef97da50e47e0e57d7ad8a6a75"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.550762 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"535e69e3ae75ca0a5e674f2e4572e8e2d849d2b68c48eedafc6639600381b7b7"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.550774 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"2f88484b900ec4569eb6e6260810a9fd3cf94e212d22f858cff90e980825845d"} Oct 02 07:34:04 crc kubenswrapper[4829]: I1002 07:34:04.550786 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"cbb2df2a7a67011f4cd0372f9a8e6fb5693ce635126454e6cf75680d9710b9e5"} Oct 02 07:34:05 crc kubenswrapper[4829]: I1002 07:34:05.559980 4829 generic.go:334] "Generic (PLEG): container finished" podID="0f1fa854-ffef-4cfe-a58f-7fc7f382af72" containerID="51fae9a9d6dde21c2c6b451d316fee6b2f43669963e47a89ea91d44ec5510306" exitCode=0 Oct 02 07:34:05 crc kubenswrapper[4829]: I1002 07:34:05.560082 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-676c-account-create-6wfbq" event={"ID":"0f1fa854-ffef-4cfe-a58f-7fc7f382af72","Type":"ContainerDied","Data":"51fae9a9d6dde21c2c6b451d316fee6b2f43669963e47a89ea91d44ec5510306"} Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.316855 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.323743 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b5txm" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.386817 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szcvc\" (UniqueName: \"kubernetes.io/projected/0f1fa854-ffef-4cfe-a58f-7fc7f382af72-kube-api-access-szcvc\") pod \"0f1fa854-ffef-4cfe-a58f-7fc7f382af72\" (UID: \"0f1fa854-ffef-4cfe-a58f-7fc7f382af72\") " Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.386887 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-combined-ca-bundle\") pod \"97b06d08-abdf-4107-9561-9b81768e183c\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.386905 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-config-data\") pod \"97b06d08-abdf-4107-9561-9b81768e183c\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.386960 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwjnz\" (UniqueName: \"kubernetes.io/projected/97b06d08-abdf-4107-9561-9b81768e183c-kube-api-access-vwjnz\") pod \"97b06d08-abdf-4107-9561-9b81768e183c\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.387056 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-db-sync-config-data\") pod \"97b06d08-abdf-4107-9561-9b81768e183c\" (UID: \"97b06d08-abdf-4107-9561-9b81768e183c\") " Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.393540 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97b06d08-abdf-4107-9561-9b81768e183c-kube-api-access-vwjnz" (OuterVolumeSpecName: "kube-api-access-vwjnz") pod "97b06d08-abdf-4107-9561-9b81768e183c" (UID: "97b06d08-abdf-4107-9561-9b81768e183c"). InnerVolumeSpecName "kube-api-access-vwjnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.394632 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "97b06d08-abdf-4107-9561-9b81768e183c" (UID: "97b06d08-abdf-4107-9561-9b81768e183c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.398073 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1fa854-ffef-4cfe-a58f-7fc7f382af72-kube-api-access-szcvc" (OuterVolumeSpecName: "kube-api-access-szcvc") pod "0f1fa854-ffef-4cfe-a58f-7fc7f382af72" (UID: "0f1fa854-ffef-4cfe-a58f-7fc7f382af72"). InnerVolumeSpecName "kube-api-access-szcvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.411479 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97b06d08-abdf-4107-9561-9b81768e183c" (UID: "97b06d08-abdf-4107-9561-9b81768e183c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.433239 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-config-data" (OuterVolumeSpecName: "config-data") pod "97b06d08-abdf-4107-9561-9b81768e183c" (UID: "97b06d08-abdf-4107-9561-9b81768e183c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.492608 4829 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.492673 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szcvc\" (UniqueName: \"kubernetes.io/projected/0f1fa854-ffef-4cfe-a58f-7fc7f382af72-kube-api-access-szcvc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.492691 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.492705 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b06d08-abdf-4107-9561-9b81768e183c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.492717 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwjnz\" (UniqueName: \"kubernetes.io/projected/97b06d08-abdf-4107-9561-9b81768e183c-kube-api-access-vwjnz\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.582099 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-b5txm" event={"ID":"97b06d08-abdf-4107-9561-9b81768e183c","Type":"ContainerDied","Data":"6b8411efbe15d97335a6c2ade94d2b0c01ff730ce21f1b476ec3ee18f8b315a8"} Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.582145 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b8411efbe15d97335a6c2ade94d2b0c01ff730ce21f1b476ec3ee18f8b315a8" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.582184 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-b5txm" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.584607 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-676c-account-create-6wfbq" event={"ID":"0f1fa854-ffef-4cfe-a58f-7fc7f382af72","Type":"ContainerDied","Data":"6ebe6f1994fb90f8a63ba2a0c0fa64f976ba9ae9e75606d3ff5d4b22c00fdaa6"} Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.584661 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ebe6f1994fb90f8a63ba2a0c0fa64f976ba9ae9e75606d3ff5d4b22c00fdaa6" Oct 02 07:34:07 crc kubenswrapper[4829]: I1002 07:34:07.584645 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-676c-account-create-6wfbq" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.181511 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.307270 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7r89\" (UniqueName: \"kubernetes.io/projected/ebc9e693-09fe-4577-a32b-cad2af43ab8a-kube-api-access-t7r89\") pod \"ebc9e693-09fe-4577-a32b-cad2af43ab8a\" (UID: \"ebc9e693-09fe-4577-a32b-cad2af43ab8a\") " Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.311995 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebc9e693-09fe-4577-a32b-cad2af43ab8a-kube-api-access-t7r89" (OuterVolumeSpecName: "kube-api-access-t7r89") pod "ebc9e693-09fe-4577-a32b-cad2af43ab8a" (UID: "ebc9e693-09fe-4577-a32b-cad2af43ab8a"). InnerVolumeSpecName "kube-api-access-t7r89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.409118 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7r89\" (UniqueName: \"kubernetes.io/projected/ebc9e693-09fe-4577-a32b-cad2af43ab8a-kube-api-access-t7r89\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.646702 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-3056-account-create-gdp9h" event={"ID":"ebc9e693-09fe-4577-a32b-cad2af43ab8a","Type":"ContainerDied","Data":"2712aab3cd8be7b517d9e50645921c7bc1e041ab53b042500606df59afd75198"} Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.646740 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2712aab3cd8be7b517d9e50645921c7bc1e041ab53b042500606df59afd75198" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.646800 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-3056-account-create-gdp9h" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.695957 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.718044 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.782275 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-grxfl"] Oct 02 07:34:08 crc kubenswrapper[4829]: E1002 07:34:08.783140 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97b06d08-abdf-4107-9561-9b81768e183c" containerName="glance-db-sync" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.783157 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="97b06d08-abdf-4107-9561-9b81768e183c" containerName="glance-db-sync" Oct 02 07:34:08 crc kubenswrapper[4829]: E1002 07:34:08.783167 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1fa854-ffef-4cfe-a58f-7fc7f382af72" containerName="mariadb-account-create" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.783173 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1fa854-ffef-4cfe-a58f-7fc7f382af72" containerName="mariadb-account-create" Oct 02 07:34:08 crc kubenswrapper[4829]: E1002 07:34:08.783188 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebc9e693-09fe-4577-a32b-cad2af43ab8a" containerName="mariadb-account-create" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.783194 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebc9e693-09fe-4577-a32b-cad2af43ab8a" containerName="mariadb-account-create" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.783378 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="97b06d08-abdf-4107-9561-9b81768e183c" containerName="glance-db-sync" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.783412 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1fa854-ffef-4cfe-a58f-7fc7f382af72" containerName="mariadb-account-create" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.783426 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebc9e693-09fe-4577-a32b-cad2af43ab8a" containerName="mariadb-account-create" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.784355 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.793041 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-grxfl"] Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.934909 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-config\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.935010 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kvlv\" (UniqueName: \"kubernetes.io/projected/0b771adf-80af-4920-806a-0d9396d52685-kube-api-access-8kvlv\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.935038 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.935088 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:08 crc kubenswrapper[4829]: I1002 07:34:08.935378 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.037430 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.037488 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-config\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.037530 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kvlv\" (UniqueName: \"kubernetes.io/projected/0b771adf-80af-4920-806a-0d9396d52685-kube-api-access-8kvlv\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.037554 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.037597 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.038505 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.038525 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.039050 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-config\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.039609 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.057937 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kvlv\" (UniqueName: \"kubernetes.io/projected/0b771adf-80af-4920-806a-0d9396d52685-kube-api-access-8kvlv\") pod \"dnsmasq-dns-5b946c75cc-grxfl\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.108912 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:09 crc kubenswrapper[4829]: I1002 07:34:09.666168 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.129579 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8fcd-account-create-7fhc4"] Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.131267 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.133758 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.148816 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8fcd-account-create-7fhc4"] Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.332420 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlj9z\" (UniqueName: \"kubernetes.io/projected/971bceb7-6e70-4502-b0bc-f5071f3998ea-kube-api-access-zlj9z\") pod \"cinder-8fcd-account-create-7fhc4\" (UID: \"971bceb7-6e70-4502-b0bc-f5071f3998ea\") " pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.434387 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlj9z\" (UniqueName: \"kubernetes.io/projected/971bceb7-6e70-4502-b0bc-f5071f3998ea-kube-api-access-zlj9z\") pod \"cinder-8fcd-account-create-7fhc4\" (UID: \"971bceb7-6e70-4502-b0bc-f5071f3998ea\") " pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.456325 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlj9z\" (UniqueName: \"kubernetes.io/projected/971bceb7-6e70-4502-b0bc-f5071f3998ea-kube-api-access-zlj9z\") pod \"cinder-8fcd-account-create-7fhc4\" (UID: \"971bceb7-6e70-4502-b0bc-f5071f3998ea\") " pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:13 crc kubenswrapper[4829]: I1002 07:34:13.460129 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:16 crc kubenswrapper[4829]: E1002 07:34:16.542015 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.65:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest" Oct 02 07:34:16 crc kubenswrapper[4829]: E1002 07:34:16.542407 4829 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.65:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest" Oct 02 07:34:16 crc kubenswrapper[4829]: E1002 07:34:16.542593 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-db-sync,Image:38.102.83.65:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/watcher/watcher.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c42qk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-db-sync-28kmg_openstack(af15cc1d-88dc-44f9-b43c-9570dfc4fd65): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:16 crc kubenswrapper[4829]: E1002 07:34:16.543802 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/watcher-db-sync-28kmg" podUID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" Oct 02 07:34:16 crc kubenswrapper[4829]: E1002 07:34:16.744007 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.65:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest\\\"\"" pod="openstack/watcher-db-sync-28kmg" podUID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.098813 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-grxfl"] Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.207043 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8fcd-account-create-7fhc4"] Oct 02 07:34:17 crc kubenswrapper[4829]: W1002 07:34:17.212696 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod971bceb7_6e70_4502_b0bc_f5071f3998ea.slice/crio-105636f85cfed77a016bf3584acacc2c71dd844426bcf4df1616d139e7cfe869 WatchSource:0}: Error finding container 105636f85cfed77a016bf3584acacc2c71dd844426bcf4df1616d139e7cfe869: Status 404 returned error can't find the container with id 105636f85cfed77a016bf3584acacc2c71dd844426bcf4df1616d139e7cfe869 Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.747983 4829 generic.go:334] "Generic (PLEG): container finished" podID="971bceb7-6e70-4502-b0bc-f5071f3998ea" containerID="a2af9b1d63b65d5d17cd15749891c841dc3fa078036044911f30035e6fbf5489" exitCode=0 Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.748068 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fcd-account-create-7fhc4" event={"ID":"971bceb7-6e70-4502-b0bc-f5071f3998ea","Type":"ContainerDied","Data":"a2af9b1d63b65d5d17cd15749891c841dc3fa078036044911f30035e6fbf5489"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.748113 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fcd-account-create-7fhc4" event={"ID":"971bceb7-6e70-4502-b0bc-f5071f3998ea","Type":"ContainerStarted","Data":"105636f85cfed77a016bf3584acacc2c71dd844426bcf4df1616d139e7cfe869"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.757381 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"046b0ea6cebff2aefa949b961997139c22127fca6b0d68f2669553a17de7eb75"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.757421 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"7ff260a90d81f7afc9d143b3625fcd9a83358cd34e86517fdffc543297406f41"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.757431 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"9da44bca-624e-49ca-8fd2-e0b0974c9ae5","Type":"ContainerStarted","Data":"15456d32eadfdb9c090b504989fa07da6e1fce4c6a3bf865cd1af6c167ea560d"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.759122 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-89hnk" event={"ID":"98649236-1856-4a42-8171-9435997ecb47","Type":"ContainerStarted","Data":"71f7a1ddbf5b094e2d3daf3cf8edf4985a804ce1b17904051424458e37f4321c"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.761192 4829 generic.go:334] "Generic (PLEG): container finished" podID="0b771adf-80af-4920-806a-0d9396d52685" containerID="f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25" exitCode=0 Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.761241 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" event={"ID":"0b771adf-80af-4920-806a-0d9396d52685","Type":"ContainerDied","Data":"f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.761260 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" event={"ID":"0b771adf-80af-4920-806a-0d9396d52685","Type":"ContainerStarted","Data":"cecb20c9575bfbe55833ec67b3c9732a28ca264ca7ecd3ab8069e5945b1c1915"} Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.784457 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-89hnk" podStartSLOduration=3.197811763 podStartE2EDuration="18.784439744s" podCreationTimestamp="2025-10-02 07:33:59 +0000 UTC" firstStartedPulling="2025-10-02 07:34:00.941144753 +0000 UTC m=+1032.280793158" lastFinishedPulling="2025-10-02 07:34:16.527772694 +0000 UTC m=+1047.867421139" observedRunningTime="2025-10-02 07:34:17.779127472 +0000 UTC m=+1049.118775877" watchObservedRunningTime="2025-10-02 07:34:17.784439744 +0000 UTC m=+1049.124088169" Oct 02 07:34:17 crc kubenswrapper[4829]: I1002 07:34:17.873211 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=53.695213956 podStartE2EDuration="1m1.873193768s" podCreationTimestamp="2025-10-02 07:33:16 +0000 UTC" firstStartedPulling="2025-10-02 07:33:55.228461894 +0000 UTC m=+1026.568110299" lastFinishedPulling="2025-10-02 07:34:03.406441706 +0000 UTC m=+1034.746090111" observedRunningTime="2025-10-02 07:34:17.872060771 +0000 UTC m=+1049.211709186" watchObservedRunningTime="2025-10-02 07:34:17.873193768 +0000 UTC m=+1049.212842173" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.116702 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-grxfl"] Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.161978 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-kr4tm"] Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.163634 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.165313 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.181606 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-kr4tm"] Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.329453 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.329974 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-config\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.330013 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.330115 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wxbt\" (UniqueName: \"kubernetes.io/projected/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-kube-api-access-9wxbt\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.330162 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.330452 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.432785 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.432856 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.432965 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-config\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.432996 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.433041 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wxbt\" (UniqueName: \"kubernetes.io/projected/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-kube-api-access-9wxbt\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.433072 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.433761 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-config\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.433816 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.433971 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.434376 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.434671 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.462850 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wxbt\" (UniqueName: \"kubernetes.io/projected/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-kube-api-access-9wxbt\") pod \"dnsmasq-dns-74f6bcbc87-kr4tm\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.481519 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.771472 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" event={"ID":"0b771adf-80af-4920-806a-0d9396d52685","Type":"ContainerStarted","Data":"9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e"} Oct 02 07:34:18 crc kubenswrapper[4829]: I1002 07:34:18.797987 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" podStartSLOduration=10.797964977 podStartE2EDuration="10.797964977s" podCreationTimestamp="2025-10-02 07:34:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:18.78564317 +0000 UTC m=+1050.125291575" watchObservedRunningTime="2025-10-02 07:34:18.797964977 +0000 UTC m=+1050.137613392" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:18.999887 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-kr4tm"] Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.110051 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.198501 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.348257 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlj9z\" (UniqueName: \"kubernetes.io/projected/971bceb7-6e70-4502-b0bc-f5071f3998ea-kube-api-access-zlj9z\") pod \"971bceb7-6e70-4502-b0bc-f5071f3998ea\" (UID: \"971bceb7-6e70-4502-b0bc-f5071f3998ea\") " Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.352330 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/971bceb7-6e70-4502-b0bc-f5071f3998ea-kube-api-access-zlj9z" (OuterVolumeSpecName: "kube-api-access-zlj9z") pod "971bceb7-6e70-4502-b0bc-f5071f3998ea" (UID: "971bceb7-6e70-4502-b0bc-f5071f3998ea"). InnerVolumeSpecName "kube-api-access-zlj9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.450482 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlj9z\" (UniqueName: \"kubernetes.io/projected/971bceb7-6e70-4502-b0bc-f5071f3998ea-kube-api-access-zlj9z\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.786079 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8fcd-account-create-7fhc4" event={"ID":"971bceb7-6e70-4502-b0bc-f5071f3998ea","Type":"ContainerDied","Data":"105636f85cfed77a016bf3584acacc2c71dd844426bcf4df1616d139e7cfe869"} Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.786122 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="105636f85cfed77a016bf3584acacc2c71dd844426bcf4df1616d139e7cfe869" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.786090 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8fcd-account-create-7fhc4" Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.788548 4829 generic.go:334] "Generic (PLEG): container finished" podID="98649236-1856-4a42-8171-9435997ecb47" containerID="71f7a1ddbf5b094e2d3daf3cf8edf4985a804ce1b17904051424458e37f4321c" exitCode=0 Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.788599 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-89hnk" event={"ID":"98649236-1856-4a42-8171-9435997ecb47","Type":"ContainerDied","Data":"71f7a1ddbf5b094e2d3daf3cf8edf4985a804ce1b17904051424458e37f4321c"} Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.792006 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" event={"ID":"e60ffb6c-bc27-44d4-9c9d-9a47227b4531","Type":"ContainerDied","Data":"741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a"} Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.792138 4829 generic.go:334] "Generic (PLEG): container finished" podID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerID="741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a" exitCode=0 Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.792210 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" event={"ID":"e60ffb6c-bc27-44d4-9c9d-9a47227b4531","Type":"ContainerStarted","Data":"3bfa825a010f641415c24cad8921471a4a6a981ca27350d578ece076931bc9b3"} Oct 02 07:34:19 crc kubenswrapper[4829]: I1002 07:34:19.792516 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" podUID="0b771adf-80af-4920-806a-0d9396d52685" containerName="dnsmasq-dns" containerID="cri-o://9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e" gracePeriod=10 Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.178736 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.263601 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-config\") pod \"0b771adf-80af-4920-806a-0d9396d52685\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.263854 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-sb\") pod \"0b771adf-80af-4920-806a-0d9396d52685\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.263907 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-dns-svc\") pod \"0b771adf-80af-4920-806a-0d9396d52685\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.263965 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kvlv\" (UniqueName: \"kubernetes.io/projected/0b771adf-80af-4920-806a-0d9396d52685-kube-api-access-8kvlv\") pod \"0b771adf-80af-4920-806a-0d9396d52685\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.264003 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-nb\") pod \"0b771adf-80af-4920-806a-0d9396d52685\" (UID: \"0b771adf-80af-4920-806a-0d9396d52685\") " Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.269069 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b771adf-80af-4920-806a-0d9396d52685-kube-api-access-8kvlv" (OuterVolumeSpecName: "kube-api-access-8kvlv") pod "0b771adf-80af-4920-806a-0d9396d52685" (UID: "0b771adf-80af-4920-806a-0d9396d52685"). InnerVolumeSpecName "kube-api-access-8kvlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.308215 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b771adf-80af-4920-806a-0d9396d52685" (UID: "0b771adf-80af-4920-806a-0d9396d52685"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.313896 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-config" (OuterVolumeSpecName: "config") pod "0b771adf-80af-4920-806a-0d9396d52685" (UID: "0b771adf-80af-4920-806a-0d9396d52685"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.316001 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0b771adf-80af-4920-806a-0d9396d52685" (UID: "0b771adf-80af-4920-806a-0d9396d52685"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.334915 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b771adf-80af-4920-806a-0d9396d52685" (UID: "0b771adf-80af-4920-806a-0d9396d52685"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.365645 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.365680 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.365695 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kvlv\" (UniqueName: \"kubernetes.io/projected/0b771adf-80af-4920-806a-0d9396d52685-kube-api-access-8kvlv\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.365710 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.365718 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b771adf-80af-4920-806a-0d9396d52685-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.806357 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" event={"ID":"e60ffb6c-bc27-44d4-9c9d-9a47227b4531","Type":"ContainerStarted","Data":"5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc"} Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.806722 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.808770 4829 generic.go:334] "Generic (PLEG): container finished" podID="0b771adf-80af-4920-806a-0d9396d52685" containerID="9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e" exitCode=0 Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.808979 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.808996 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" event={"ID":"0b771adf-80af-4920-806a-0d9396d52685","Type":"ContainerDied","Data":"9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e"} Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.809191 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-grxfl" event={"ID":"0b771adf-80af-4920-806a-0d9396d52685","Type":"ContainerDied","Data":"cecb20c9575bfbe55833ec67b3c9732a28ca264ca7ecd3ab8069e5945b1c1915"} Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.809258 4829 scope.go:117] "RemoveContainer" containerID="9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.849998 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" podStartSLOduration=2.849976789 podStartE2EDuration="2.849976789s" podCreationTimestamp="2025-10-02 07:34:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:20.8459495 +0000 UTC m=+1052.185597945" watchObservedRunningTime="2025-10-02 07:34:20.849976789 +0000 UTC m=+1052.189625204" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.867199 4829 scope.go:117] "RemoveContainer" containerID="f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.912064 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-grxfl"] Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.921359 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-grxfl"] Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.934814 4829 scope.go:117] "RemoveContainer" containerID="9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e" Oct 02 07:34:20 crc kubenswrapper[4829]: E1002 07:34:20.935635 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e\": container with ID starting with 9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e not found: ID does not exist" containerID="9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.935708 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e"} err="failed to get container status \"9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e\": rpc error: code = NotFound desc = could not find container \"9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e\": container with ID starting with 9aa6d0e3c9805e156e90984f4f0ade84211c03ecedd4792ef89d4685ca078a7e not found: ID does not exist" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.935750 4829 scope.go:117] "RemoveContainer" containerID="f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25" Oct 02 07:34:20 crc kubenswrapper[4829]: E1002 07:34:20.936722 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25\": container with ID starting with f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25 not found: ID does not exist" containerID="f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25" Oct 02 07:34:20 crc kubenswrapper[4829]: I1002 07:34:20.936804 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25"} err="failed to get container status \"f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25\": rpc error: code = NotFound desc = could not find container \"f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25\": container with ID starting with f8054bc8a836e382fa442c6c974555db73f580dce4253c88006042f684b92f25 not found: ID does not exist" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.281606 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.389015 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-combined-ca-bundle\") pod \"98649236-1856-4a42-8171-9435997ecb47\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.389346 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkqkm\" (UniqueName: \"kubernetes.io/projected/98649236-1856-4a42-8171-9435997ecb47-kube-api-access-zkqkm\") pod \"98649236-1856-4a42-8171-9435997ecb47\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.389486 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-config-data\") pod \"98649236-1856-4a42-8171-9435997ecb47\" (UID: \"98649236-1856-4a42-8171-9435997ecb47\") " Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.397248 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98649236-1856-4a42-8171-9435997ecb47-kube-api-access-zkqkm" (OuterVolumeSpecName: "kube-api-access-zkqkm") pod "98649236-1856-4a42-8171-9435997ecb47" (UID: "98649236-1856-4a42-8171-9435997ecb47"). InnerVolumeSpecName "kube-api-access-zkqkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.415680 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98649236-1856-4a42-8171-9435997ecb47" (UID: "98649236-1856-4a42-8171-9435997ecb47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.455489 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-config-data" (OuterVolumeSpecName: "config-data") pod "98649236-1856-4a42-8171-9435997ecb47" (UID: "98649236-1856-4a42-8171-9435997ecb47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.479469 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b771adf-80af-4920-806a-0d9396d52685" path="/var/lib/kubelet/pods/0b771adf-80af-4920-806a-0d9396d52685/volumes" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.492380 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.492429 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkqkm\" (UniqueName: \"kubernetes.io/projected/98649236-1856-4a42-8171-9435997ecb47-kube-api-access-zkqkm\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.492452 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98649236-1856-4a42-8171-9435997ecb47-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.822367 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-89hnk" event={"ID":"98649236-1856-4a42-8171-9435997ecb47","Type":"ContainerDied","Data":"b6570d3ee095dd877ce2a17f923412f5dd8a0c9441c3ffd26302870fcd20af31"} Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.822425 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6570d3ee095dd877ce2a17f923412f5dd8a0c9441c3ffd26302870fcd20af31" Oct 02 07:34:21 crc kubenswrapper[4829]: I1002 07:34:21.822445 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-89hnk" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.141503 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-qtxdz"] Oct 02 07:34:22 crc kubenswrapper[4829]: E1002 07:34:22.142200 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b771adf-80af-4920-806a-0d9396d52685" containerName="init" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.142721 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b771adf-80af-4920-806a-0d9396d52685" containerName="init" Oct 02 07:34:22 crc kubenswrapper[4829]: E1002 07:34:22.142747 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b771adf-80af-4920-806a-0d9396d52685" containerName="dnsmasq-dns" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.142755 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b771adf-80af-4920-806a-0d9396d52685" containerName="dnsmasq-dns" Oct 02 07:34:22 crc kubenswrapper[4829]: E1002 07:34:22.142783 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98649236-1856-4a42-8171-9435997ecb47" containerName="keystone-db-sync" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.142792 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="98649236-1856-4a42-8171-9435997ecb47" containerName="keystone-db-sync" Oct 02 07:34:22 crc kubenswrapper[4829]: E1002 07:34:22.142820 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="971bceb7-6e70-4502-b0bc-f5071f3998ea" containerName="mariadb-account-create" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.142826 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="971bceb7-6e70-4502-b0bc-f5071f3998ea" containerName="mariadb-account-create" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.143042 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b771adf-80af-4920-806a-0d9396d52685" containerName="dnsmasq-dns" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.143064 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="98649236-1856-4a42-8171-9435997ecb47" containerName="keystone-db-sync" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.143082 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="971bceb7-6e70-4502-b0bc-f5071f3998ea" containerName="mariadb-account-create" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.144062 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.147844 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.148043 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.148088 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxrff" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.151457 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.163404 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-kr4tm"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.169829 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qtxdz"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.193848 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-nml6m"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.195706 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.246187 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-nml6m"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315751 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-scripts\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315831 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-credential-keys\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315874 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trtfx\" (UniqueName: \"kubernetes.io/projected/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-kube-api-access-trtfx\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315930 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-combined-ca-bundle\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315948 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315977 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-config\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.315997 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9bm7\" (UniqueName: \"kubernetes.io/projected/36189b22-d628-455e-a107-86debfbb86d5-kube-api-access-t9bm7\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.316022 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.316054 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.316083 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-config-data\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.316103 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-fernet-keys\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.316128 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.417359 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-scripts\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.417997 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-credential-keys\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418043 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trtfx\" (UniqueName: \"kubernetes.io/projected/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-kube-api-access-trtfx\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418081 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418098 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-combined-ca-bundle\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418123 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-config\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418165 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9bm7\" (UniqueName: \"kubernetes.io/projected/36189b22-d628-455e-a107-86debfbb86d5-kube-api-access-t9bm7\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418196 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418253 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418293 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-config-data\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418315 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-fernet-keys\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.418343 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.419113 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-config\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.419121 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.419159 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.419409 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-svc\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.419721 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.421848 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-config-data\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.422735 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-credential-keys\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.426834 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-scripts\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.432799 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-57686b68f9-mfp5f"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.433990 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-combined-ca-bundle\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.434296 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.441564 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-fernet-keys\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.462437 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57686b68f9-mfp5f"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.476434 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.476690 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.476873 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-qwfvf" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.477550 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.483971 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9bm7\" (UniqueName: \"kubernetes.io/projected/36189b22-d628-455e-a107-86debfbb86d5-kube-api-access-t9bm7\") pod \"keystone-bootstrap-qtxdz\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.489884 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trtfx\" (UniqueName: \"kubernetes.io/projected/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-kube-api-access-trtfx\") pod \"dnsmasq-dns-847c4cc679-nml6m\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.513940 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.531058 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.546115 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.562966 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.563133 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.586351 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-tzrbj"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.587578 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.610291 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.610566 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.610867 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.611071 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ssb9v" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.627455 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e8ba59d-d691-47df-848b-f208dfe5fffb-logs\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.627567 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-556zd\" (UniqueName: \"kubernetes.io/projected/2e8ba59d-d691-47df-848b-f208dfe5fffb-kube-api-access-556zd\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.627611 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-scripts\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.639194 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e8ba59d-d691-47df-848b-f208dfe5fffb-horizon-secret-key\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.639326 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-config-data\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.678645 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-tzrbj"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.741065 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-config-data\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742136 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-scripts\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742198 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742255 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742299 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e8ba59d-d691-47df-848b-f208dfe5fffb-logs\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742350 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-556zd\" (UniqueName: \"kubernetes.io/projected/2e8ba59d-d691-47df-848b-f208dfe5fffb-kube-api-access-556zd\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742419 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmw6f\" (UniqueName: \"kubernetes.io/projected/8bd0e965-1cf0-4edd-856e-d424c91d80a9-kube-api-access-qmw6f\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742476 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-combined-ca-bundle\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742512 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-scripts\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742533 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-config-data\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.742962 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e8ba59d-d691-47df-848b-f208dfe5fffb-logs\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.743080 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8jtd\" (UniqueName: \"kubernetes.io/projected/e302e0a1-26ca-4e34-8edb-9c26df3f1812-kube-api-access-d8jtd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.743114 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-config\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.743167 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-run-httpd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.743218 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e8ba59d-d691-47df-848b-f208dfe5fffb-horizon-secret-key\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.746959 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-log-httpd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.751254 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-scripts\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.752210 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-config-data\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.760139 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.764375 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e8ba59d-d691-47df-848b-f208dfe5fffb-horizon-secret-key\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.774195 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dtlnv"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.776026 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.780673 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pv8ch" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.791556 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.798183 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-556zd\" (UniqueName: \"kubernetes.io/projected/2e8ba59d-d691-47df-848b-f208dfe5fffb-kube-api-access-556zd\") pod \"horizon-57686b68f9-mfp5f\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.837401 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-8tjd9"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.838466 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.894704 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-run-httpd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.894904 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-log-httpd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.894978 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-scripts\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.895081 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.897004 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.897303 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmw6f\" (UniqueName: \"kubernetes.io/projected/8bd0e965-1cf0-4edd-856e-d424c91d80a9-kube-api-access-qmw6f\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.897366 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-combined-ca-bundle\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.897419 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-config-data\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.897538 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8jtd\" (UniqueName: \"kubernetes.io/projected/e302e0a1-26ca-4e34-8edb-9c26df3f1812-kube-api-access-d8jtd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.897709 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-config\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.904048 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-log-httpd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.905957 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerName="dnsmasq-dns" containerID="cri-o://5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc" gracePeriod=10 Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.906026 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.906153 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.906882 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-combined-ca-bundle\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.907205 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.907238 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zfpss" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.909981 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-run-httpd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.925327 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.934365 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-scripts\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.936260 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-config-data\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.941371 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-config\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.941792 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dtlnv"] Oct 02 07:34:22 crc kubenswrapper[4829]: I1002 07:34:22.979078 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.003778 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8tjd9"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.013396 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmw6f\" (UniqueName: \"kubernetes.io/projected/8bd0e965-1cf0-4edd-856e-d424c91d80a9-kube-api-access-qmw6f\") pod \"neutron-db-sync-tzrbj\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.013545 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8jtd\" (UniqueName: \"kubernetes.io/projected/e302e0a1-26ca-4e34-8edb-9c26df3f1812-kube-api-access-d8jtd\") pod \"ceilometer-0\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " pod="openstack/ceilometer-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.021816 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.027809 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031337 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-db-sync-config-data\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031383 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-logs\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031419 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r8xs\" (UniqueName: \"kubernetes.io/projected/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-kube-api-access-9r8xs\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031455 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-combined-ca-bundle\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031500 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-combined-ca-bundle\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031532 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-config-data\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031556 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkwhv\" (UniqueName: \"kubernetes.io/projected/1b3627e8-07df-4d85-a723-a763eb14c52b-kube-api-access-lkwhv\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.031616 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-scripts\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.084836 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-nml6m"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135668 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-combined-ca-bundle\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135725 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-config-data\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135751 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkwhv\" (UniqueName: \"kubernetes.io/projected/1b3627e8-07df-4d85-a723-a763eb14c52b-kube-api-access-lkwhv\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135804 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-scripts\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135849 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-db-sync-config-data\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135878 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-logs\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135920 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r8xs\" (UniqueName: \"kubernetes.io/projected/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-kube-api-access-9r8xs\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.135948 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-combined-ca-bundle\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.136639 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-logs\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.140008 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b86bd75fc-6wpv2"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.142084 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.146004 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-config-data\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.153053 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-db-sync-config-data\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.156519 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-combined-ca-bundle\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.175445 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-scripts\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.175794 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-combined-ca-bundle\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.184786 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r8xs\" (UniqueName: \"kubernetes.io/projected/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-kube-api-access-9r8xs\") pod \"placement-db-sync-8tjd9\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.193825 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkwhv\" (UniqueName: \"kubernetes.io/projected/1b3627e8-07df-4d85-a723-a763eb14c52b-kube-api-access-lkwhv\") pod \"barbican-db-sync-dtlnv\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.200832 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.202443 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.206558 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gtlmv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.206845 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.207028 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.207150 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.243270 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.259298 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.274435 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5j67n"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.275900 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.295908 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b86bd75fc-6wpv2"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.302210 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5j67n"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.311932 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.313588 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.315331 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.320186 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.320526 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.327305 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.339823 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.339883 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.339914 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/40d40c68-19b2-4d91-834b-d54ffde9806d-horizon-secret-key\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.339930 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40d40c68-19b2-4d91-834b-d54ffde9806d-logs\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.339946 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-logs\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.339989 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-config-data\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340004 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340022 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340043 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340061 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-scripts\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340105 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340125 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4x4p\" (UniqueName: \"kubernetes.io/projected/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-kube-api-access-g4x4p\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.340148 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtptc\" (UniqueName: \"kubernetes.io/projected/40d40c68-19b2-4d91-834b-d54ffde9806d-kube-api-access-gtptc\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.441875 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-config-data\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444421 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444463 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444489 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444861 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444902 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444925 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-scripts\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444944 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq257\" (UniqueName: \"kubernetes.io/projected/f65a4787-32b1-4883-b6b3-895e412001fb-kube-api-access-gq257\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.444978 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448457 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448511 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448590 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-config\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448659 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448694 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4x4p\" (UniqueName: \"kubernetes.io/projected/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-kube-api-access-g4x4p\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtptc\" (UniqueName: \"kubernetes.io/projected/40d40c68-19b2-4d91-834b-d54ffde9806d-kube-api-access-gtptc\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448795 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448843 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448879 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448936 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.448987 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449037 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/40d40c68-19b2-4d91-834b-d54ffde9806d-horizon-secret-key\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449192 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40d40c68-19b2-4d91-834b-d54ffde9806d-logs\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449244 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-logs\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449303 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449340 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449406 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.443118 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-config-data\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449422 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.449680 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.452510 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-scripts\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.454142 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40d40c68-19b2-4d91-834b-d54ffde9806d-logs\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.454273 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr7sw\" (UniqueName: \"kubernetes.io/projected/bd90d04d-dd93-423e-96ce-2534e46f079a-kube-api-access-sr7sw\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.454638 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-logs\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.455835 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-scripts\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.457170 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.462283 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.468317 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/40d40c68-19b2-4d91-834b-d54ffde9806d-horizon-secret-key\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.473498 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtptc\" (UniqueName: \"kubernetes.io/projected/40d40c68-19b2-4d91-834b-d54ffde9806d-kube-api-access-gtptc\") pod \"horizon-7b86bd75fc-6wpv2\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.473589 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-config-data\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.474900 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4x4p\" (UniqueName: \"kubernetes.io/projected/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-kube-api-access-g4x4p\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.485523 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.506361 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xhn5v"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.507355 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xhn5v"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.507430 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.512809 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.512953 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4dmnc" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.514471 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.542356 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.549477 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.557948 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558025 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558062 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558138 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558162 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558205 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558241 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr7sw\" (UniqueName: \"kubernetes.io/projected/bd90d04d-dd93-423e-96ce-2534e46f079a-kube-api-access-sr7sw\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558270 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558291 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558314 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq257\" (UniqueName: \"kubernetes.io/projected/f65a4787-32b1-4883-b6b3-895e412001fb-kube-api-access-gq257\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558334 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558369 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558392 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558416 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-config\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.558913 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.560834 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.561140 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.561182 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.561385 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.561639 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-config\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.561752 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.564333 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.566332 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.568663 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.571286 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.597090 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.611443 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-nml6m"] Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.613358 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq257\" (UniqueName: \"kubernetes.io/projected/f65a4787-32b1-4883-b6b3-895e412001fb-kube-api-access-gq257\") pod \"dnsmasq-dns-785d8bcb8c-5j67n\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.614556 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr7sw\" (UniqueName: \"kubernetes.io/projected/bd90d04d-dd93-423e-96ce-2534e46f079a-kube-api-access-sr7sw\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.628898 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.642136 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.659795 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-scripts\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.659901 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea5c0100-6e79-43fb-9761-62c7e758e891-etc-machine-id\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.659946 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-combined-ca-bundle\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.659969 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-config-data\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.659996 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-db-sync-config-data\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.660061 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x57f7\" (UniqueName: \"kubernetes.io/projected/ea5c0100-6e79-43fb-9761-62c7e758e891-kube-api-access-x57f7\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.761218 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-scripts\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.763623 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea5c0100-6e79-43fb-9761-62c7e758e891-etc-machine-id\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.763664 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-combined-ca-bundle\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.763701 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-config-data\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.763736 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-db-sync-config-data\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.763780 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x57f7\" (UniqueName: \"kubernetes.io/projected/ea5c0100-6e79-43fb-9761-62c7e758e891-kube-api-access-x57f7\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.766187 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea5c0100-6e79-43fb-9761-62c7e758e891-etc-machine-id\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.794916 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-scripts\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.797627 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-db-sync-config-data\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.798710 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-config-data\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.802611 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x57f7\" (UniqueName: \"kubernetes.io/projected/ea5c0100-6e79-43fb-9761-62c7e758e891-kube-api-access-x57f7\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.803071 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-combined-ca-bundle\") pod \"cinder-db-sync-xhn5v\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.900080 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.922099 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.926970 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.959545 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" event={"ID":"39e0108c-9b79-4af0-a2b1-8155fdb67ee7","Type":"ContainerStarted","Data":"904fe74999296a9a77e2d53911270a19287ab947204993ecd8cec104e3fbc205"} Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.963312 4829 generic.go:334] "Generic (PLEG): container finished" podID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerID="5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc" exitCode=0 Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.963352 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" event={"ID":"e60ffb6c-bc27-44d4-9c9d-9a47227b4531","Type":"ContainerDied","Data":"5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc"} Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.963377 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" event={"ID":"e60ffb6c-bc27-44d4-9c9d-9a47227b4531","Type":"ContainerDied","Data":"3bfa825a010f641415c24cad8921471a4a6a981ca27350d578ece076931bc9b3"} Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.963426 4829 scope.go:117] "RemoveContainer" containerID="5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc" Oct 02 07:34:23 crc kubenswrapper[4829]: I1002 07:34:23.963579 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-kr4tm" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.037001 4829 scope.go:117] "RemoveContainer" containerID="741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.051768 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-qtxdz"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.068232 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-nb\") pod \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.068301 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-config\") pod \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.068336 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wxbt\" (UniqueName: \"kubernetes.io/projected/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-kube-api-access-9wxbt\") pod \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.068388 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-sb\") pod \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.068466 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-svc\") pod \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.068548 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-swift-storage-0\") pod \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\" (UID: \"e60ffb6c-bc27-44d4-9c9d-9a47227b4531\") " Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.097665 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.098935 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-kube-api-access-9wxbt" (OuterVolumeSpecName: "kube-api-access-9wxbt") pod "e60ffb6c-bc27-44d4-9c9d-9a47227b4531" (UID: "e60ffb6c-bc27-44d4-9c9d-9a47227b4531"). InnerVolumeSpecName "kube-api-access-9wxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4829]: W1002 07:34:24.115114 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode302e0a1_26ca_4e34_8edb_9c26df3f1812.slice/crio-458fbbb2ee00e56a2eac486eff2fba6e6c15a408ee78771c85bffe90504ef928 WatchSource:0}: Error finding container 458fbbb2ee00e56a2eac486eff2fba6e6c15a408ee78771c85bffe90504ef928: Status 404 returned error can't find the container with id 458fbbb2ee00e56a2eac486eff2fba6e6c15a408ee78771c85bffe90504ef928 Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.119751 4829 scope.go:117] "RemoveContainer" containerID="5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc" Oct 02 07:34:24 crc kubenswrapper[4829]: E1002 07:34:24.125555 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc\": container with ID starting with 5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc not found: ID does not exist" containerID="5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.125594 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc"} err="failed to get container status \"5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc\": rpc error: code = NotFound desc = could not find container \"5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc\": container with ID starting with 5215d9cb8590b06b68e1e2ee91e850dd822da9f99f63d15b1a9cfdb0dcf233dc not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.125652 4829 scope.go:117] "RemoveContainer" containerID="741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a" Oct 02 07:34:24 crc kubenswrapper[4829]: E1002 07:34:24.126890 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a\": container with ID starting with 741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a not found: ID does not exist" containerID="741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.126956 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a"} err="failed to get container status \"741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a\": rpc error: code = NotFound desc = could not find container \"741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a\": container with ID starting with 741ad4deba0604975c0667561ce718994c39f2ed8de369ace04b18a3e71b9e3a not found: ID does not exist" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.170678 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wxbt\" (UniqueName: \"kubernetes.io/projected/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-kube-api-access-9wxbt\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.285398 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-tzrbj"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.329873 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-config" (OuterVolumeSpecName: "config") pod "e60ffb6c-bc27-44d4-9c9d-9a47227b4531" (UID: "e60ffb6c-bc27-44d4-9c9d-9a47227b4531"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.363945 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e60ffb6c-bc27-44d4-9c9d-9a47227b4531" (UID: "e60ffb6c-bc27-44d4-9c9d-9a47227b4531"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.368887 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e60ffb6c-bc27-44d4-9c9d-9a47227b4531" (UID: "e60ffb6c-bc27-44d4-9c9d-9a47227b4531"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.375486 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e60ffb6c-bc27-44d4-9c9d-9a47227b4531" (UID: "e60ffb6c-bc27-44d4-9c9d-9a47227b4531"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.379751 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.379778 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.379811 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.379820 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.387567 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.404510 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e60ffb6c-bc27-44d4-9c9d-9a47227b4531" (UID: "e60ffb6c-bc27-44d4-9c9d-9a47227b4531"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.482463 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57686b68f9-mfp5f"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.483658 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e60ffb6c-bc27-44d4-9c9d-9a47227b4531-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.485569 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-cfb69cc5c-vm8vd"] Oct 02 07:34:24 crc kubenswrapper[4829]: E1002 07:34:24.485929 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerName="dnsmasq-dns" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.485945 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerName="dnsmasq-dns" Oct 02 07:34:24 crc kubenswrapper[4829]: E1002 07:34:24.485959 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerName="init" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.485966 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerName="init" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.486154 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" containerName="dnsmasq-dns" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.488116 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.580155 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cfb69cc5c-vm8vd"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.596797 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.597435 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3069acac-b31f-4fa1-b902-023722aa791a-logs\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.598443 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3069acac-b31f-4fa1-b902-023722aa791a-horizon-secret-key\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.598704 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t4ql\" (UniqueName: \"kubernetes.io/projected/3069acac-b31f-4fa1-b902-023722aa791a-kube-api-access-5t4ql\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.599299 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-scripts\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.599395 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-config-data\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.614492 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dtlnv"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.642326 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57686b68f9-mfp5f"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.654567 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-kr4tm"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.661057 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-kr4tm"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.700994 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-scripts\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.701373 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-config-data\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.701449 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3069acac-b31f-4fa1-b902-023722aa791a-logs\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.701491 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3069acac-b31f-4fa1-b902-023722aa791a-horizon-secret-key\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.701537 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t4ql\" (UniqueName: \"kubernetes.io/projected/3069acac-b31f-4fa1-b902-023722aa791a-kube-api-access-5t4ql\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.702925 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-config-data\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.703462 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-scripts\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.709630 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3069acac-b31f-4fa1-b902-023722aa791a-logs\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.726440 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t4ql\" (UniqueName: \"kubernetes.io/projected/3069acac-b31f-4fa1-b902-023722aa791a-kube-api-access-5t4ql\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.737264 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3069acac-b31f-4fa1-b902-023722aa791a-horizon-secret-key\") pod \"horizon-cfb69cc5c-vm8vd\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.766972 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b86bd75fc-6wpv2"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.789884 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8tjd9"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.797045 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.854497 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5j67n"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.865172 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.880369 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xhn5v"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.887167 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:24 crc kubenswrapper[4829]: W1002 07:34:24.940420 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea5c0100_6e79_43fb_9761_62c7e758e891.slice/crio-6650b173eabc531f2dec4d9ded5eeb40e22f13dda30a8751d0be6b48b1ad8efb WatchSource:0}: Error finding container 6650b173eabc531f2dec4d9ded5eeb40e22f13dda30a8751d0be6b48b1ad8efb: Status 404 returned error can't find the container with id 6650b173eabc531f2dec4d9ded5eeb40e22f13dda30a8751d0be6b48b1ad8efb Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.957789 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.974750 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dtlnv" event={"ID":"1b3627e8-07df-4d85-a723-a763eb14c52b","Type":"ContainerStarted","Data":"45a2e7265e54e482cec6e946c0e1ff8372fdcdf6b661e4150c0c14a9edc9147c"} Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.978238 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tzrbj" event={"ID":"8bd0e965-1cf0-4edd-856e-d424c91d80a9","Type":"ContainerStarted","Data":"1aa8fb8ff73f092bc301d8f6acdd4500b47f9ce3b4bdc9d83539574ff1a72b76"} Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.978293 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tzrbj" event={"ID":"8bd0e965-1cf0-4edd-856e-d424c91d80a9","Type":"ContainerStarted","Data":"5b07f43b3008fa6048f5c0647d98347a192e221b1e0bf0d07fb82bead9e10570"} Oct 02 07:34:24 crc kubenswrapper[4829]: I1002 07:34:24.985920 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b86bd75fc-6wpv2" event={"ID":"40d40c68-19b2-4d91-834b-d54ffde9806d","Type":"ContainerStarted","Data":"1c5716bccf7ecd8424bd2cefa4098b0909ec58625d06f0e037d6e2a73a452280"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.001921 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd90d04d-dd93-423e-96ce-2534e46f079a","Type":"ContainerStarted","Data":"89bfd5f0f033c843cde2d4b21ea677035490ab3a11eb7f50843310c5072ff598"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.013302 4829 generic.go:334] "Generic (PLEG): container finished" podID="39e0108c-9b79-4af0-a2b1-8155fdb67ee7" containerID="3e1f7a3a46d55995de371bb15981d93d4329006e7463884367e1e09c30cde149" exitCode=0 Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.013418 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" event={"ID":"39e0108c-9b79-4af0-a2b1-8155fdb67ee7","Type":"ContainerDied","Data":"3e1f7a3a46d55995de371bb15981d93d4329006e7463884367e1e09c30cde149"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.025666 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-tzrbj" podStartSLOduration=3.025642436 podStartE2EDuration="3.025642436s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:25.010826659 +0000 UTC m=+1056.350475064" watchObservedRunningTime="2025-10-02 07:34:25.025642436 +0000 UTC m=+1056.365290841" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.055471 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8tjd9" event={"ID":"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4","Type":"ContainerStarted","Data":"82777d608b8f25d0796f7087c28e6c8a0aa200b47bf06ab317465c159e89713e"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.063126 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xhn5v" event={"ID":"ea5c0100-6e79-43fb-9761-62c7e758e891","Type":"ContainerStarted","Data":"6650b173eabc531f2dec4d9ded5eeb40e22f13dda30a8751d0be6b48b1ad8efb"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.064643 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57686b68f9-mfp5f" event={"ID":"2e8ba59d-d691-47df-848b-f208dfe5fffb","Type":"ContainerStarted","Data":"a9abc750d6674cf89e168a75a781a222a16b3cfb60b2eebbd19d374c41266b5a"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.067786 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qtxdz" event={"ID":"36189b22-d628-455e-a107-86debfbb86d5","Type":"ContainerStarted","Data":"229bc912bc4f9bd9d62366c21ed8e572caa493456101b39333c2c03734cdb5d9"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.067808 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qtxdz" event={"ID":"36189b22-d628-455e-a107-86debfbb86d5","Type":"ContainerStarted","Data":"60756eecd6ca98db31f0cec187c5e1dba1f7b3c250c59381ec0a03422353037f"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.073178 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerStarted","Data":"458fbbb2ee00e56a2eac486eff2fba6e6c15a408ee78771c85bffe90504ef928"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.078875 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" event={"ID":"f65a4787-32b1-4883-b6b3-895e412001fb","Type":"ContainerStarted","Data":"8088103963fc9559d505cdc4dee1c867bccb596acf47b203d331ee3e7e6d1e67"} Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.105769 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-qtxdz" podStartSLOduration=3.105746131 podStartE2EDuration="3.105746131s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:25.089391913 +0000 UTC m=+1056.429040318" watchObservedRunningTime="2025-10-02 07:34:25.105746131 +0000 UTC m=+1056.445394536" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.508516 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e60ffb6c-bc27-44d4-9c9d-9a47227b4531" path="/var/lib/kubelet/pods/e60ffb6c-bc27-44d4-9c9d-9a47227b4531/volumes" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.510074 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.604920 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cfb69cc5c-vm8vd"] Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.662601 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-svc\") pod \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.663030 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-nb\") pod \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.663156 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trtfx\" (UniqueName: \"kubernetes.io/projected/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-kube-api-access-trtfx\") pod \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.663186 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-sb\") pod \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.663256 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-config\") pod \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.663291 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-swift-storage-0\") pod \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\" (UID: \"39e0108c-9b79-4af0-a2b1-8155fdb67ee7\") " Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.681554 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-kube-api-access-trtfx" (OuterVolumeSpecName: "kube-api-access-trtfx") pod "39e0108c-9b79-4af0-a2b1-8155fdb67ee7" (UID: "39e0108c-9b79-4af0-a2b1-8155fdb67ee7"). InnerVolumeSpecName "kube-api-access-trtfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.703921 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "39e0108c-9b79-4af0-a2b1-8155fdb67ee7" (UID: "39e0108c-9b79-4af0-a2b1-8155fdb67ee7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.718027 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "39e0108c-9b79-4af0-a2b1-8155fdb67ee7" (UID: "39e0108c-9b79-4af0-a2b1-8155fdb67ee7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.719773 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-config" (OuterVolumeSpecName: "config") pod "39e0108c-9b79-4af0-a2b1-8155fdb67ee7" (UID: "39e0108c-9b79-4af0-a2b1-8155fdb67ee7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.731742 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "39e0108c-9b79-4af0-a2b1-8155fdb67ee7" (UID: "39e0108c-9b79-4af0-a2b1-8155fdb67ee7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.743269 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "39e0108c-9b79-4af0-a2b1-8155fdb67ee7" (UID: "39e0108c-9b79-4af0-a2b1-8155fdb67ee7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.772689 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.772725 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trtfx\" (UniqueName: \"kubernetes.io/projected/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-kube-api-access-trtfx\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.772736 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.772746 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.772755 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:25 crc kubenswrapper[4829]: I1002 07:34:25.772777 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/39e0108c-9b79-4af0-a2b1-8155fdb67ee7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.093567 4829 generic.go:334] "Generic (PLEG): container finished" podID="f65a4787-32b1-4883-b6b3-895e412001fb" containerID="6c98ae7a568dbee9d3183312d9fce81ec583f32185de85ba77df9e19bc072a29" exitCode=0 Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.094158 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" event={"ID":"f65a4787-32b1-4883-b6b3-895e412001fb","Type":"ContainerDied","Data":"6c98ae7a568dbee9d3183312d9fce81ec583f32185de85ba77df9e19bc072a29"} Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.099487 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7","Type":"ContainerStarted","Data":"7998fa6bf413109db3f0c2e0e6dacf60e725e067e2376ab0fb8ad83edcd1c0ac"} Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.101329 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" event={"ID":"39e0108c-9b79-4af0-a2b1-8155fdb67ee7","Type":"ContainerDied","Data":"904fe74999296a9a77e2d53911270a19287ab947204993ecd8cec104e3fbc205"} Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.101365 4829 scope.go:117] "RemoveContainer" containerID="3e1f7a3a46d55995de371bb15981d93d4329006e7463884367e1e09c30cde149" Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.101477 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-nml6m" Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.118253 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cfb69cc5c-vm8vd" event={"ID":"3069acac-b31f-4fa1-b902-023722aa791a","Type":"ContainerStarted","Data":"14a12a22a2c87e3b8469939797a2ea2b231c4b614e2675f31c8b3393862f519c"} Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.209165 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-nml6m"] Oct 02 07:34:26 crc kubenswrapper[4829]: I1002 07:34:26.227598 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-nml6m"] Oct 02 07:34:26 crc kubenswrapper[4829]: E1002 07:34:26.354378 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39e0108c_9b79_4af0_a2b1_8155fdb67ee7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39e0108c_9b79_4af0_a2b1_8155fdb67ee7.slice/crio-904fe74999296a9a77e2d53911270a19287ab947204993ecd8cec104e3fbc205\": RecentStats: unable to find data in memory cache]" Oct 02 07:34:27 crc kubenswrapper[4829]: I1002 07:34:27.147725 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7","Type":"ContainerStarted","Data":"165c49e159d193cd6ecf21682947faa46dd8e76b8e7939aea820b0f079e1d6bb"} Oct 02 07:34:27 crc kubenswrapper[4829]: I1002 07:34:27.153272 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd90d04d-dd93-423e-96ce-2534e46f079a","Type":"ContainerStarted","Data":"d6b43c1d0b6db65f3dbc959e43911a9fdfe24f6fe79150deb555c3f78433adae"} Oct 02 07:34:27 crc kubenswrapper[4829]: I1002 07:34:27.157140 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" event={"ID":"f65a4787-32b1-4883-b6b3-895e412001fb","Type":"ContainerStarted","Data":"37562d498b8f6e6ff4c4bb710c27c7b0805b91e7634c26144c52d25284d34733"} Oct 02 07:34:27 crc kubenswrapper[4829]: I1002 07:34:27.158747 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:27 crc kubenswrapper[4829]: I1002 07:34:27.184257 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" podStartSLOduration=5.184241708 podStartE2EDuration="5.184241708s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:27.180652383 +0000 UTC m=+1058.520300788" watchObservedRunningTime="2025-10-02 07:34:27.184241708 +0000 UTC m=+1058.523890113" Oct 02 07:34:27 crc kubenswrapper[4829]: I1002 07:34:27.495852 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39e0108c-9b79-4af0-a2b1-8155fdb67ee7" path="/var/lib/kubelet/pods/39e0108c-9b79-4af0-a2b1-8155fdb67ee7/volumes" Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.179266 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7","Type":"ContainerStarted","Data":"f8e5b0c9a451d784ead6fb8c18667de1ee9239624091a871fd8d2c3d05805744"} Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.179338 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-log" containerID="cri-o://165c49e159d193cd6ecf21682947faa46dd8e76b8e7939aea820b0f079e1d6bb" gracePeriod=30 Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.179366 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-httpd" containerID="cri-o://f8e5b0c9a451d784ead6fb8c18667de1ee9239624091a871fd8d2c3d05805744" gracePeriod=30 Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.195605 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd90d04d-dd93-423e-96ce-2534e46f079a","Type":"ContainerStarted","Data":"5e5353c0fa828f0f6548198b70ba6d94ef1cb9a60e20605dc6b4441f04cc1971"} Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.195690 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-log" containerID="cri-o://d6b43c1d0b6db65f3dbc959e43911a9fdfe24f6fe79150deb555c3f78433adae" gracePeriod=30 Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.195740 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-httpd" containerID="cri-o://5e5353c0fa828f0f6548198b70ba6d94ef1cb9a60e20605dc6b4441f04cc1971" gracePeriod=30 Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.220187 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.220167174 podStartE2EDuration="6.220167174s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:28.209984236 +0000 UTC m=+1059.549632641" watchObservedRunningTime="2025-10-02 07:34:28.220167174 +0000 UTC m=+1059.559815579" Oct 02 07:34:28 crc kubenswrapper[4829]: I1002 07:34:28.253207 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.25317877 podStartE2EDuration="5.25317877s" podCreationTimestamp="2025-10-02 07:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:28.25197277 +0000 UTC m=+1059.591621175" watchObservedRunningTime="2025-10-02 07:34:28.25317877 +0000 UTC m=+1059.592827185" Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.215066 4829 generic.go:334] "Generic (PLEG): container finished" podID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerID="5e5353c0fa828f0f6548198b70ba6d94ef1cb9a60e20605dc6b4441f04cc1971" exitCode=0 Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.215455 4829 generic.go:334] "Generic (PLEG): container finished" podID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerID="d6b43c1d0b6db65f3dbc959e43911a9fdfe24f6fe79150deb555c3f78433adae" exitCode=143 Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.215161 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd90d04d-dd93-423e-96ce-2534e46f079a","Type":"ContainerDied","Data":"5e5353c0fa828f0f6548198b70ba6d94ef1cb9a60e20605dc6b4441f04cc1971"} Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.215513 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd90d04d-dd93-423e-96ce-2534e46f079a","Type":"ContainerDied","Data":"d6b43c1d0b6db65f3dbc959e43911a9fdfe24f6fe79150deb555c3f78433adae"} Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.222499 4829 generic.go:334] "Generic (PLEG): container finished" podID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerID="f8e5b0c9a451d784ead6fb8c18667de1ee9239624091a871fd8d2c3d05805744" exitCode=0 Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.222528 4829 generic.go:334] "Generic (PLEG): container finished" podID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerID="165c49e159d193cd6ecf21682947faa46dd8e76b8e7939aea820b0f079e1d6bb" exitCode=143 Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.222622 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7","Type":"ContainerDied","Data":"f8e5b0c9a451d784ead6fb8c18667de1ee9239624091a871fd8d2c3d05805744"} Oct 02 07:34:29 crc kubenswrapper[4829]: I1002 07:34:29.222679 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7","Type":"ContainerDied","Data":"165c49e159d193cd6ecf21682947faa46dd8e76b8e7939aea820b0f079e1d6bb"} Oct 02 07:34:30 crc kubenswrapper[4829]: I1002 07:34:30.234614 4829 generic.go:334] "Generic (PLEG): container finished" podID="36189b22-d628-455e-a107-86debfbb86d5" containerID="229bc912bc4f9bd9d62366c21ed8e572caa493456101b39333c2c03734cdb5d9" exitCode=0 Oct 02 07:34:30 crc kubenswrapper[4829]: I1002 07:34:30.234678 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qtxdz" event={"ID":"36189b22-d628-455e-a107-86debfbb86d5","Type":"ContainerDied","Data":"229bc912bc4f9bd9d62366c21ed8e572caa493456101b39333c2c03734cdb5d9"} Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.121892 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b86bd75fc-6wpv2"] Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.160136 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7dcfb8bd44-ws7cq"] Oct 02 07:34:31 crc kubenswrapper[4829]: E1002 07:34:31.161804 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e0108c-9b79-4af0-a2b1-8155fdb67ee7" containerName="init" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.161836 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e0108c-9b79-4af0-a2b1-8155fdb67ee7" containerName="init" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.162075 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e0108c-9b79-4af0-a2b1-8155fdb67ee7" containerName="init" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.163557 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.175955 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.188436 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7dcfb8bd44-ws7cq"] Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246074 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-tls-certs\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246133 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-scripts\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246193 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a176a27-612f-499a-82c4-645d9cdd0b8b-logs\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246250 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-secret-key\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246302 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-combined-ca-bundle\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246365 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlxn6\" (UniqueName: \"kubernetes.io/projected/3a176a27-612f-499a-82c4-645d9cdd0b8b-kube-api-access-jlxn6\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.246435 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-config-data\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.340400 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cfb69cc5c-vm8vd"] Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348288 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-combined-ca-bundle\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348348 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlxn6\" (UniqueName: \"kubernetes.io/projected/3a176a27-612f-499a-82c4-645d9cdd0b8b-kube-api-access-jlxn6\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348393 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-config-data\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348461 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-tls-certs\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348480 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-scripts\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348509 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a176a27-612f-499a-82c4-645d9cdd0b8b-logs\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.348524 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-secret-key\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.350180 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-scripts\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.350191 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-config-data\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.350535 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a176a27-612f-499a-82c4-645d9cdd0b8b-logs\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.357922 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-tls-certs\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.366842 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-combined-ca-bundle\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.367901 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5df99f8898-76z9s"] Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.369708 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.376119 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlxn6\" (UniqueName: \"kubernetes.io/projected/3a176a27-612f-499a-82c4-645d9cdd0b8b-kube-api-access-jlxn6\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.376490 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-secret-key\") pod \"horizon-7dcfb8bd44-ws7cq\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.376486 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5df99f8898-76z9s"] Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.505593 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.568710 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3407e6aa-fc06-4a2e-bd97-b8540ae12167-logs\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.570070 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-combined-ca-bundle\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.570325 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-horizon-secret-key\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.570435 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3407e6aa-fc06-4a2e-bd97-b8540ae12167-scripts\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.570459 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-horizon-tls-certs\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.570694 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc4k6\" (UniqueName: \"kubernetes.io/projected/3407e6aa-fc06-4a2e-bd97-b8540ae12167-kube-api-access-kc4k6\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.570756 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3407e6aa-fc06-4a2e-bd97-b8540ae12167-config-data\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672440 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3407e6aa-fc06-4a2e-bd97-b8540ae12167-scripts\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672487 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-horizon-tls-certs\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672574 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc4k6\" (UniqueName: \"kubernetes.io/projected/3407e6aa-fc06-4a2e-bd97-b8540ae12167-kube-api-access-kc4k6\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672600 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3407e6aa-fc06-4a2e-bd97-b8540ae12167-config-data\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672635 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3407e6aa-fc06-4a2e-bd97-b8540ae12167-logs\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672689 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-combined-ca-bundle\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.672745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-horizon-secret-key\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.674567 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3407e6aa-fc06-4a2e-bd97-b8540ae12167-scripts\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.674939 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3407e6aa-fc06-4a2e-bd97-b8540ae12167-logs\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.675111 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3407e6aa-fc06-4a2e-bd97-b8540ae12167-config-data\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.678477 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-combined-ca-bundle\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.687258 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-horizon-tls-certs\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.689655 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc4k6\" (UniqueName: \"kubernetes.io/projected/3407e6aa-fc06-4a2e-bd97-b8540ae12167-kube-api-access-kc4k6\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.689935 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3407e6aa-fc06-4a2e-bd97-b8540ae12167-horizon-secret-key\") pod \"horizon-5df99f8898-76z9s\" (UID: \"3407e6aa-fc06-4a2e-bd97-b8540ae12167\") " pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:31 crc kubenswrapper[4829]: I1002 07:34:31.759949 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.506746 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.513659 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586481 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-config-data\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586740 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-scripts\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586765 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-httpd-run\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586780 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-config-data\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586799 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-combined-ca-bundle\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586823 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-combined-ca-bundle\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586864 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-internal-tls-certs\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586887 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-httpd-run\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586913 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-scripts\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586931 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586969 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-logs\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.586987 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-logs\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.587005 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr7sw\" (UniqueName: \"kubernetes.io/projected/bd90d04d-dd93-423e-96ce-2534e46f079a-kube-api-access-sr7sw\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.587026 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-public-tls-certs\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.587062 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4x4p\" (UniqueName: \"kubernetes.io/projected/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-kube-api-access-g4x4p\") pod \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\" (UID: \"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.587092 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"bd90d04d-dd93-423e-96ce-2534e46f079a\" (UID: \"bd90d04d-dd93-423e-96ce-2534e46f079a\") " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.588354 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.589869 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-logs" (OuterVolumeSpecName: "logs") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.590418 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-logs" (OuterVolumeSpecName: "logs") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.590215 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.595137 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-scripts" (OuterVolumeSpecName: "scripts") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.596030 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.601059 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd90d04d-dd93-423e-96ce-2534e46f079a-kube-api-access-sr7sw" (OuterVolumeSpecName: "kube-api-access-sr7sw") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "kube-api-access-sr7sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.602296 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.602818 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-kube-api-access-g4x4p" (OuterVolumeSpecName: "kube-api-access-g4x4p") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "kube-api-access-g4x4p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.617501 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-scripts" (OuterVolumeSpecName: "scripts") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688363 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688392 4829 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688402 4829 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688428 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688437 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688445 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd90d04d-dd93-423e-96ce-2534e46f079a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688453 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr7sw\" (UniqueName: \"kubernetes.io/projected/bd90d04d-dd93-423e-96ce-2534e46f079a-kube-api-access-sr7sw\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688464 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688472 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4x4p\" (UniqueName: \"kubernetes.io/projected/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-kube-api-access-g4x4p\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.688486 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.689497 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.717323 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.717417 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.733769 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.749627 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-config-data" (OuterVolumeSpecName: "config-data") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.749714 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.773396 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-config-data" (OuterVolumeSpecName: "config-data") pod "bd90d04d-dd93-423e-96ce-2534e46f079a" (UID: "bd90d04d-dd93-423e-96ce-2534e46f079a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.777828 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" (UID: "cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793649 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793683 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793693 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793706 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793715 4829 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd90d04d-dd93-423e-96ce-2534e46f079a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793724 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793736 4829 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:32 crc kubenswrapper[4829]: I1002 07:34:32.793750 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.280848 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7","Type":"ContainerDied","Data":"7998fa6bf413109db3f0c2e0e6dacf60e725e067e2376ab0fb8ad83edcd1c0ac"} Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.280864 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.280909 4829 scope.go:117] "RemoveContainer" containerID="f8e5b0c9a451d784ead6fb8c18667de1ee9239624091a871fd8d2c3d05805744" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.286493 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd90d04d-dd93-423e-96ce-2534e46f079a","Type":"ContainerDied","Data":"89bfd5f0f033c843cde2d4b21ea677035490ab3a11eb7f50843310c5072ff598"} Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.286563 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.317558 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.332384 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.336289 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.348118 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.377892 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: E1002 07:34:33.384540 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-log" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.384571 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-log" Oct 02 07:34:33 crc kubenswrapper[4829]: E1002 07:34:33.384590 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-log" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.384597 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-log" Oct 02 07:34:33 crc kubenswrapper[4829]: E1002 07:34:33.384609 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-httpd" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.384616 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-httpd" Oct 02 07:34:33 crc kubenswrapper[4829]: E1002 07:34:33.384638 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-httpd" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.384644 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-httpd" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.386054 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-log" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.386078 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-log" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.386088 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" containerName="glance-httpd" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.386101 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" containerName="glance-httpd" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.387119 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.389057 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.389254 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.389393 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.389548 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.389565 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gtlmv" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.391203 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.392710 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.393572 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.396549 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.412107 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.488886 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd90d04d-dd93-423e-96ce-2534e46f079a" path="/var/lib/kubelet/pods/bd90d04d-dd93-423e-96ce-2534e46f079a/volumes" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.489671 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7" path="/var/lib/kubelet/pods/cc3a0d0a-d486-41b9-918a-f4c6d9c38fe7/volumes" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.517731 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518253 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518402 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518460 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518501 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shlds\" (UniqueName: \"kubernetes.io/projected/63630a30-0c56-4bdb-b259-984b81733bcc-kube-api-access-shlds\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518534 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518571 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-config-data\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518595 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-logs\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518611 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518625 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-scripts\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518646 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-logs\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518668 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518730 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.518753 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.519044 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.519395 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw8lj\" (UniqueName: \"kubernetes.io/projected/8104fa43-766f-4747-bfde-8efb050e4629-kube-api-access-nw8lj\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.623367 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.623434 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.623772 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.623810 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.623857 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.624083 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shlds\" (UniqueName: \"kubernetes.io/projected/63630a30-0c56-4bdb-b259-984b81733bcc-kube-api-access-shlds\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.624203 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.624787 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-config-data\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.624816 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-logs\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.624971 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.625091 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-scripts\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.625120 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-logs\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.625140 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.625394 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.625396 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.625415 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.624895 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.626845 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-logs\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.627045 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-logs\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.627571 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.628756 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.628833 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw8lj\" (UniqueName: \"kubernetes.io/projected/8104fa43-766f-4747-bfde-8efb050e4629-kube-api-access-nw8lj\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.629847 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-scripts\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.630152 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.630465 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-config-data\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.630749 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.632400 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.642965 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.644202 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.644947 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.646555 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw8lj\" (UniqueName: \"kubernetes.io/projected/8104fa43-766f-4747-bfde-8efb050e4629-kube-api-access-nw8lj\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.648933 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shlds\" (UniqueName: \"kubernetes.io/projected/63630a30-0c56-4bdb-b259-984b81733bcc-kube-api-access-shlds\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.662083 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.670049 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.708523 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.716911 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.901628 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.985332 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kj4jn"] Oct 02 07:34:33 crc kubenswrapper[4829]: I1002 07:34:33.985547 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-kj4jn" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" containerID="cri-o://6f4ebf4a95d5f6ba9071dd198b392585ee5ec237cba3ceb3d9ba50b2058a7128" gracePeriod=10 Oct 02 07:34:34 crc kubenswrapper[4829]: I1002 07:34:34.303557 4829 generic.go:334] "Generic (PLEG): container finished" podID="5919e153-3626-4764-8110-ae849b21ff17" containerID="6f4ebf4a95d5f6ba9071dd198b392585ee5ec237cba3ceb3d9ba50b2058a7128" exitCode=0 Oct 02 07:34:34 crc kubenswrapper[4829]: I1002 07:34:34.303634 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kj4jn" event={"ID":"5919e153-3626-4764-8110-ae849b21ff17","Type":"ContainerDied","Data":"6f4ebf4a95d5f6ba9071dd198b392585ee5ec237cba3ceb3d9ba50b2058a7128"} Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.818722 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.976934 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-combined-ca-bundle\") pod \"36189b22-d628-455e-a107-86debfbb86d5\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.977487 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-config-data\") pod \"36189b22-d628-455e-a107-86debfbb86d5\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.977612 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-scripts\") pod \"36189b22-d628-455e-a107-86debfbb86d5\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.977683 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-credential-keys\") pod \"36189b22-d628-455e-a107-86debfbb86d5\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.977735 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9bm7\" (UniqueName: \"kubernetes.io/projected/36189b22-d628-455e-a107-86debfbb86d5-kube-api-access-t9bm7\") pod \"36189b22-d628-455e-a107-86debfbb86d5\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.977777 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-fernet-keys\") pod \"36189b22-d628-455e-a107-86debfbb86d5\" (UID: \"36189b22-d628-455e-a107-86debfbb86d5\") " Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.986174 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "36189b22-d628-455e-a107-86debfbb86d5" (UID: "36189b22-d628-455e-a107-86debfbb86d5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.986308 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "36189b22-d628-455e-a107-86debfbb86d5" (UID: "36189b22-d628-455e-a107-86debfbb86d5"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.987652 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36189b22-d628-455e-a107-86debfbb86d5-kube-api-access-t9bm7" (OuterVolumeSpecName: "kube-api-access-t9bm7") pod "36189b22-d628-455e-a107-86debfbb86d5" (UID: "36189b22-d628-455e-a107-86debfbb86d5"). InnerVolumeSpecName "kube-api-access-t9bm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:35 crc kubenswrapper[4829]: I1002 07:34:35.988526 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-scripts" (OuterVolumeSpecName: "scripts") pod "36189b22-d628-455e-a107-86debfbb86d5" (UID: "36189b22-d628-455e-a107-86debfbb86d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.019622 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-config-data" (OuterVolumeSpecName: "config-data") pod "36189b22-d628-455e-a107-86debfbb86d5" (UID: "36189b22-d628-455e-a107-86debfbb86d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.024695 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36189b22-d628-455e-a107-86debfbb86d5" (UID: "36189b22-d628-455e-a107-86debfbb86d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.081690 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.081729 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.081796 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.081811 4829 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.081824 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9bm7\" (UniqueName: \"kubernetes.io/projected/36189b22-d628-455e-a107-86debfbb86d5-kube-api-access-t9bm7\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.081836 4829 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36189b22-d628-455e-a107-86debfbb86d5-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.327253 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-qtxdz" event={"ID":"36189b22-d628-455e-a107-86debfbb86d5","Type":"ContainerDied","Data":"60756eecd6ca98db31f0cec187c5e1dba1f7b3c250c59381ec0a03422353037f"} Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.327288 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60756eecd6ca98db31f0cec187c5e1dba1f7b3c250c59381ec0a03422353037f" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.327340 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-qtxdz" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.907477 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-qtxdz"] Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.910311 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-kj4jn" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Oct 02 07:34:36 crc kubenswrapper[4829]: I1002 07:34:36.915717 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-qtxdz"] Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.010707 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-g5dzf"] Oct 02 07:34:37 crc kubenswrapper[4829]: E1002 07:34:37.011434 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36189b22-d628-455e-a107-86debfbb86d5" containerName="keystone-bootstrap" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.011468 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="36189b22-d628-455e-a107-86debfbb86d5" containerName="keystone-bootstrap" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.011790 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="36189b22-d628-455e-a107-86debfbb86d5" containerName="keystone-bootstrap" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.012873 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.015208 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.015297 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxrff" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.015469 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.015622 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.022385 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g5dzf"] Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.103008 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-scripts\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.103063 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-fernet-keys\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.103108 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-credential-keys\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.103134 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpqsr\" (UniqueName: \"kubernetes.io/projected/e43b663f-cfca-459a-b513-ed16fdecd2f6-kube-api-access-tpqsr\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.103375 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-combined-ca-bundle\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.103420 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-config-data\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.205322 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-combined-ca-bundle\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.205723 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-config-data\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.205817 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-scripts\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.205842 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-fernet-keys\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.205886 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-credential-keys\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.205913 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpqsr\" (UniqueName: \"kubernetes.io/projected/e43b663f-cfca-459a-b513-ed16fdecd2f6-kube-api-access-tpqsr\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.211087 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-scripts\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.211513 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-fernet-keys\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.215371 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-config-data\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.218535 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-credential-keys\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.225361 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-combined-ca-bundle\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.226347 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpqsr\" (UniqueName: \"kubernetes.io/projected/e43b663f-cfca-459a-b513-ed16fdecd2f6-kube-api-access-tpqsr\") pod \"keystone-bootstrap-g5dzf\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.338047 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:37 crc kubenswrapper[4829]: I1002 07:34:37.481657 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36189b22-d628-455e-a107-86debfbb86d5" path="/var/lib/kubelet/pods/36189b22-d628-455e-a107-86debfbb86d5/volumes" Oct 02 07:34:40 crc kubenswrapper[4829]: E1002 07:34:40.436351 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 02 07:34:40 crc kubenswrapper[4829]: E1002 07:34:40.436913 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n667hcfh5dhc5h584h55ch4hfbhd4hcfh679h54h7dh57hfbh596h644h6fh676h68fhb7h66h56dh8chcfh5bh565h598h9fhfhcbh89q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gtptc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7b86bd75fc-6wpv2_openstack(40d40c68-19b2-4d91-834b-d54ffde9806d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:40 crc kubenswrapper[4829]: E1002 07:34:40.438865 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7b86bd75fc-6wpv2" podUID="40d40c68-19b2-4d91-834b-d54ffde9806d" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.855750 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.856314 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9r8xs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-8tjd9_openstack(8adee8c4-a6f1-42cb-ae32-a7cc57b911b4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.857500 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-8tjd9" podUID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.891086 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.891286 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n58hd4h696h64bh564h8fhc7h58fh56fhb6h5c6h98h574h57h658hd6h5c8h59fhdfh94hc6h5dh6hd7hbdh574h656h5fch665h667h558h564q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5t4ql,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-cfb69cc5c-vm8vd_openstack(3069acac-b31f-4fa1-b902-023722aa791a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.893449 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-cfb69cc5c-vm8vd" podUID="3069acac-b31f-4fa1-b902-023722aa791a" Oct 02 07:34:41 crc kubenswrapper[4829]: I1002 07:34:41.910074 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-kj4jn" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.913419 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.913665 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n575h55bhb8h585h94hc7hffh7fh6dh694hffh5f8h586h599hf8h598h5b8h7bh5b4h55bh5d6h576h66fh5ch645h58ch579h649hc5h66h68ch58dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-556zd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-57686b68f9-mfp5f_openstack(2e8ba59d-d691-47df-848b-f208dfe5fffb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:41 crc kubenswrapper[4829]: E1002 07:34:41.916828 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-57686b68f9-mfp5f" podUID="2e8ba59d-d691-47df-848b-f208dfe5fffb" Oct 02 07:34:42 crc kubenswrapper[4829]: I1002 07:34:42.393981 4829 generic.go:334] "Generic (PLEG): container finished" podID="8bd0e965-1cf0-4edd-856e-d424c91d80a9" containerID="1aa8fb8ff73f092bc301d8f6acdd4500b47f9ce3b4bdc9d83539574ff1a72b76" exitCode=0 Oct 02 07:34:42 crc kubenswrapper[4829]: I1002 07:34:42.394013 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tzrbj" event={"ID":"8bd0e965-1cf0-4edd-856e-d424c91d80a9","Type":"ContainerDied","Data":"1aa8fb8ff73f092bc301d8f6acdd4500b47f9ce3b4bdc9d83539574ff1a72b76"} Oct 02 07:34:42 crc kubenswrapper[4829]: E1002 07:34:42.396419 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-8tjd9" podUID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" Oct 02 07:34:49 crc kubenswrapper[4829]: E1002 07:34:49.605208 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 02 07:34:49 crc kubenswrapper[4829]: E1002 07:34:49.605719 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lkwhv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-dtlnv_openstack(1b3627e8-07df-4d85-a723-a763eb14c52b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:49 crc kubenswrapper[4829]: E1002 07:34:49.606929 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-dtlnv" podUID="1b3627e8-07df-4d85-a723-a763eb14c52b" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.677078 4829 scope.go:117] "RemoveContainer" containerID="165c49e159d193cd6ecf21682947faa46dd8e76b8e7939aea820b0f079e1d6bb" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.803011 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.812649 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.816106 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.829820 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.832978 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.878199 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-config-data\") pod \"40d40c68-19b2-4d91-834b-d54ffde9806d\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.878285 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-scripts\") pod \"40d40c68-19b2-4d91-834b-d54ffde9806d\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.878371 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40d40c68-19b2-4d91-834b-d54ffde9806d-logs\") pod \"40d40c68-19b2-4d91-834b-d54ffde9806d\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.878415 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtptc\" (UniqueName: \"kubernetes.io/projected/40d40c68-19b2-4d91-834b-d54ffde9806d-kube-api-access-gtptc\") pod \"40d40c68-19b2-4d91-834b-d54ffde9806d\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.878506 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/40d40c68-19b2-4d91-834b-d54ffde9806d-horizon-secret-key\") pod \"40d40c68-19b2-4d91-834b-d54ffde9806d\" (UID: \"40d40c68-19b2-4d91-834b-d54ffde9806d\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.878741 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40d40c68-19b2-4d91-834b-d54ffde9806d-logs" (OuterVolumeSpecName: "logs") pod "40d40c68-19b2-4d91-834b-d54ffde9806d" (UID: "40d40c68-19b2-4d91-834b-d54ffde9806d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.879137 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-config-data" (OuterVolumeSpecName: "config-data") pod "40d40c68-19b2-4d91-834b-d54ffde9806d" (UID: "40d40c68-19b2-4d91-834b-d54ffde9806d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.879288 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-scripts" (OuterVolumeSpecName: "scripts") pod "40d40c68-19b2-4d91-834b-d54ffde9806d" (UID: "40d40c68-19b2-4d91-834b-d54ffde9806d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.883446 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40d40c68-19b2-4d91-834b-d54ffde9806d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "40d40c68-19b2-4d91-834b-d54ffde9806d" (UID: "40d40c68-19b2-4d91-834b-d54ffde9806d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.887399 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d40c68-19b2-4d91-834b-d54ffde9806d-kube-api-access-gtptc" (OuterVolumeSpecName: "kube-api-access-gtptc") pod "40d40c68-19b2-4d91-834b-d54ffde9806d" (UID: "40d40c68-19b2-4d91-834b-d54ffde9806d"). InnerVolumeSpecName "kube-api-access-gtptc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979406 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-combined-ca-bundle\") pod \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979481 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-dns-svc\") pod \"5919e153-3626-4764-8110-ae849b21ff17\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979542 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3069acac-b31f-4fa1-b902-023722aa791a-horizon-secret-key\") pod \"3069acac-b31f-4fa1-b902-023722aa791a\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979581 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5dsl\" (UniqueName: \"kubernetes.io/projected/5919e153-3626-4764-8110-ae849b21ff17-kube-api-access-w5dsl\") pod \"5919e153-3626-4764-8110-ae849b21ff17\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979618 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-sb\") pod \"5919e153-3626-4764-8110-ae849b21ff17\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979640 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-config-data\") pod \"3069acac-b31f-4fa1-b902-023722aa791a\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979668 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmw6f\" (UniqueName: \"kubernetes.io/projected/8bd0e965-1cf0-4edd-856e-d424c91d80a9-kube-api-access-qmw6f\") pod \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979713 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-config\") pod \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\" (UID: \"8bd0e965-1cf0-4edd-856e-d424c91d80a9\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979773 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-config\") pod \"5919e153-3626-4764-8110-ae849b21ff17\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979800 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-556zd\" (UniqueName: \"kubernetes.io/projected/2e8ba59d-d691-47df-848b-f208dfe5fffb-kube-api-access-556zd\") pod \"2e8ba59d-d691-47df-848b-f208dfe5fffb\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979841 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3069acac-b31f-4fa1-b902-023722aa791a-logs\") pod \"3069acac-b31f-4fa1-b902-023722aa791a\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979866 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e8ba59d-d691-47df-848b-f208dfe5fffb-logs\") pod \"2e8ba59d-d691-47df-848b-f208dfe5fffb\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979905 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-scripts\") pod \"2e8ba59d-d691-47df-848b-f208dfe5fffb\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979953 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-config-data\") pod \"2e8ba59d-d691-47df-848b-f208dfe5fffb\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.979976 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e8ba59d-d691-47df-848b-f208dfe5fffb-horizon-secret-key\") pod \"2e8ba59d-d691-47df-848b-f208dfe5fffb\" (UID: \"2e8ba59d-d691-47df-848b-f208dfe5fffb\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980005 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-scripts\") pod \"3069acac-b31f-4fa1-b902-023722aa791a\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980039 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-nb\") pod \"5919e153-3626-4764-8110-ae849b21ff17\" (UID: \"5919e153-3626-4764-8110-ae849b21ff17\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980079 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t4ql\" (UniqueName: \"kubernetes.io/projected/3069acac-b31f-4fa1-b902-023722aa791a-kube-api-access-5t4ql\") pod \"3069acac-b31f-4fa1-b902-023722aa791a\" (UID: \"3069acac-b31f-4fa1-b902-023722aa791a\") " Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980077 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3069acac-b31f-4fa1-b902-023722aa791a-logs" (OuterVolumeSpecName: "logs") pod "3069acac-b31f-4fa1-b902-023722aa791a" (UID: "3069acac-b31f-4fa1-b902-023722aa791a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980558 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtptc\" (UniqueName: \"kubernetes.io/projected/40d40c68-19b2-4d91-834b-d54ffde9806d-kube-api-access-gtptc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980579 4829 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/40d40c68-19b2-4d91-834b-d54ffde9806d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980594 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980608 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40d40c68-19b2-4d91-834b-d54ffde9806d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980622 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3069acac-b31f-4fa1-b902-023722aa791a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980634 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40d40c68-19b2-4d91-834b-d54ffde9806d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980592 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-scripts" (OuterVolumeSpecName: "scripts") pod "3069acac-b31f-4fa1-b902-023722aa791a" (UID: "3069acac-b31f-4fa1-b902-023722aa791a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.980744 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-config-data" (OuterVolumeSpecName: "config-data") pod "2e8ba59d-d691-47df-848b-f208dfe5fffb" (UID: "2e8ba59d-d691-47df-848b-f208dfe5fffb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.981270 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e8ba59d-d691-47df-848b-f208dfe5fffb-logs" (OuterVolumeSpecName: "logs") pod "2e8ba59d-d691-47df-848b-f208dfe5fffb" (UID: "2e8ba59d-d691-47df-848b-f208dfe5fffb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.981374 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-config-data" (OuterVolumeSpecName: "config-data") pod "3069acac-b31f-4fa1-b902-023722aa791a" (UID: "3069acac-b31f-4fa1-b902-023722aa791a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.981489 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-scripts" (OuterVolumeSpecName: "scripts") pod "2e8ba59d-d691-47df-848b-f208dfe5fffb" (UID: "2e8ba59d-d691-47df-848b-f208dfe5fffb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.983630 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e8ba59d-d691-47df-848b-f208dfe5fffb-kube-api-access-556zd" (OuterVolumeSpecName: "kube-api-access-556zd") pod "2e8ba59d-d691-47df-848b-f208dfe5fffb" (UID: "2e8ba59d-d691-47df-848b-f208dfe5fffb"). InnerVolumeSpecName "kube-api-access-556zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.983993 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3069acac-b31f-4fa1-b902-023722aa791a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3069acac-b31f-4fa1-b902-023722aa791a" (UID: "3069acac-b31f-4fa1-b902-023722aa791a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.986001 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3069acac-b31f-4fa1-b902-023722aa791a-kube-api-access-5t4ql" (OuterVolumeSpecName: "kube-api-access-5t4ql") pod "3069acac-b31f-4fa1-b902-023722aa791a" (UID: "3069acac-b31f-4fa1-b902-023722aa791a"). InnerVolumeSpecName "kube-api-access-5t4ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.986427 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e8ba59d-d691-47df-848b-f208dfe5fffb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "2e8ba59d-d691-47df-848b-f208dfe5fffb" (UID: "2e8ba59d-d691-47df-848b-f208dfe5fffb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.987352 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5919e153-3626-4764-8110-ae849b21ff17-kube-api-access-w5dsl" (OuterVolumeSpecName: "kube-api-access-w5dsl") pod "5919e153-3626-4764-8110-ae849b21ff17" (UID: "5919e153-3626-4764-8110-ae849b21ff17"). InnerVolumeSpecName "kube-api-access-w5dsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:49 crc kubenswrapper[4829]: I1002 07:34:49.988998 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bd0e965-1cf0-4edd-856e-d424c91d80a9-kube-api-access-qmw6f" (OuterVolumeSpecName: "kube-api-access-qmw6f") pod "8bd0e965-1cf0-4edd-856e-d424c91d80a9" (UID: "8bd0e965-1cf0-4edd-856e-d424c91d80a9"). InnerVolumeSpecName "kube-api-access-qmw6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.006884 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bd0e965-1cf0-4edd-856e-d424c91d80a9" (UID: "8bd0e965-1cf0-4edd-856e-d424c91d80a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.009173 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-config" (OuterVolumeSpecName: "config") pod "8bd0e965-1cf0-4edd-856e-d424c91d80a9" (UID: "8bd0e965-1cf0-4edd-856e-d424c91d80a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.029870 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5919e153-3626-4764-8110-ae849b21ff17" (UID: "5919e153-3626-4764-8110-ae849b21ff17"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.034656 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-config" (OuterVolumeSpecName: "config") pod "5919e153-3626-4764-8110-ae849b21ff17" (UID: "5919e153-3626-4764-8110-ae849b21ff17"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.037473 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5919e153-3626-4764-8110-ae849b21ff17" (UID: "5919e153-3626-4764-8110-ae849b21ff17"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.045339 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5919e153-3626-4764-8110-ae849b21ff17" (UID: "5919e153-3626-4764-8110-ae849b21ff17"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.054121 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5df99f8898-76z9s"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082525 4829 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3069acac-b31f-4fa1-b902-023722aa791a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082575 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5dsl\" (UniqueName: \"kubernetes.io/projected/5919e153-3626-4764-8110-ae849b21ff17-kube-api-access-w5dsl\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082593 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082606 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082622 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmw6f\" (UniqueName: \"kubernetes.io/projected/8bd0e965-1cf0-4edd-856e-d424c91d80a9-kube-api-access-qmw6f\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082635 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082646 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082658 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-556zd\" (UniqueName: \"kubernetes.io/projected/2e8ba59d-d691-47df-848b-f208dfe5fffb-kube-api-access-556zd\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082669 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e8ba59d-d691-47df-848b-f208dfe5fffb-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082680 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082691 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2e8ba59d-d691-47df-848b-f208dfe5fffb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082702 4829 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2e8ba59d-d691-47df-848b-f208dfe5fffb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082713 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3069acac-b31f-4fa1-b902-023722aa791a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082726 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082738 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t4ql\" (UniqueName: \"kubernetes.io/projected/3069acac-b31f-4fa1-b902-023722aa791a-kube-api-access-5t4ql\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082778 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bd0e965-1cf0-4edd-856e-d424c91d80a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.082790 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5919e153-3626-4764-8110-ae849b21ff17-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.469459 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b86bd75fc-6wpv2" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.469472 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b86bd75fc-6wpv2" event={"ID":"40d40c68-19b2-4d91-834b-d54ffde9806d","Type":"ContainerDied","Data":"1c5716bccf7ecd8424bd2cefa4098b0909ec58625d06f0e037d6e2a73a452280"} Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.472916 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cfb69cc5c-vm8vd" event={"ID":"3069acac-b31f-4fa1-b902-023722aa791a","Type":"ContainerDied","Data":"14a12a22a2c87e3b8469939797a2ea2b231c4b614e2675f31c8b3393862f519c"} Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.472954 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cfb69cc5c-vm8vd" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.475847 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kj4jn" event={"ID":"5919e153-3626-4764-8110-ae849b21ff17","Type":"ContainerDied","Data":"7b5071013f9d8a59e3b0b9433a27e679ce6ae456aaf1cea747161df2a21e0cea"} Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.475915 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kj4jn" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.476815 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57686b68f9-mfp5f" event={"ID":"2e8ba59d-d691-47df-848b-f208dfe5fffb","Type":"ContainerDied","Data":"a9abc750d6674cf89e168a75a781a222a16b3cfb60b2eebbd19d374c41266b5a"} Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.476890 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57686b68f9-mfp5f" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.482177 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-tzrbj" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.483918 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-tzrbj" event={"ID":"8bd0e965-1cf0-4edd-856e-d424c91d80a9","Type":"ContainerDied","Data":"5b07f43b3008fa6048f5c0647d98347a192e221b1e0bf0d07fb82bead9e10570"} Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.483964 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b07f43b3008fa6048f5c0647d98347a192e221b1e0bf0d07fb82bead9e10570" Oct 02 07:34:50 crc kubenswrapper[4829]: E1002 07:34:50.488723 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-dtlnv" podUID="1b3627e8-07df-4d85-a723-a763eb14c52b" Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.559231 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57686b68f9-mfp5f"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.567467 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-57686b68f9-mfp5f"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.639941 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b86bd75fc-6wpv2"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.648882 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b86bd75fc-6wpv2"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.658405 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kj4jn"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.669852 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kj4jn"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.694482 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cfb69cc5c-vm8vd"] Oct 02 07:34:50 crc kubenswrapper[4829]: I1002 07:34:50.701718 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-cfb69cc5c-vm8vd"] Oct 02 07:34:50 crc kubenswrapper[4829]: W1002 07:34:50.955957 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3407e6aa_fc06_4a2e_bd97_b8540ae12167.slice/crio-9585517a3e9e205d346954ccacb0c4ce059338dd017dc51cc71ce46f45fc2f15 WatchSource:0}: Error finding container 9585517a3e9e205d346954ccacb0c4ce059338dd017dc51cc71ce46f45fc2f15: Status 404 returned error can't find the container with id 9585517a3e9e205d346954ccacb0c4ce059338dd017dc51cc71ce46f45fc2f15 Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.034109 4829 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.034284 4829 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x57f7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xhn5v_openstack(ea5c0100-6e79-43fb-9761-62c7e758e891): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.038068 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xhn5v" podUID="ea5c0100-6e79-43fb-9761-62c7e758e891" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.039696 4829 scope.go:117] "RemoveContainer" containerID="5e5353c0fa828f0f6548198b70ba6d94ef1cb9a60e20605dc6b4441f04cc1971" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.075281 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-9g6ss"] Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.075681 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bd0e965-1cf0-4edd-856e-d424c91d80a9" containerName="neutron-db-sync" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.075693 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bd0e965-1cf0-4edd-856e-d424c91d80a9" containerName="neutron-db-sync" Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.075716 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="init" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.075722 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="init" Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.075732 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.075738 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.075916 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bd0e965-1cf0-4edd-856e-d424c91d80a9" containerName="neutron-db-sync" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.075939 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.076927 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.081978 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-9g6ss"] Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.139515 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66cb4fcc98-c5psq"] Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.142529 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.146694 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-ssb9v" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.146875 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.148104 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.148137 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.153827 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66cb4fcc98-c5psq"] Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.193651 4829 scope.go:117] "RemoveContainer" containerID="d6b43c1d0b6db65f3dbc959e43911a9fdfe24f6fe79150deb555c3f78433adae" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.205812 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.205859 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.205964 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6stp4\" (UniqueName: \"kubernetes.io/projected/a5724411-e0dd-4ab9-a898-0ecfbfe64385-kube-api-access-6stp4\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206065 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-config\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206151 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-ovndb-tls-certs\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206410 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-combined-ca-bundle\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206501 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206559 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-httpd-config\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206584 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwphc\" (UniqueName: \"kubernetes.io/projected/5052ce64-e1d7-429f-9256-f004ef04bc5a-kube-api-access-hwphc\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206850 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-config\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.206915 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.243405 4829 scope.go:117] "RemoveContainer" containerID="6f4ebf4a95d5f6ba9071dd198b392585ee5ec237cba3ceb3d9ba50b2058a7128" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.295684 4829 scope.go:117] "RemoveContainer" containerID="66fd2865c1c504ddac447ea792768342a37d7a32f1d50b45059da5b204438574" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.308806 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.308894 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-httpd-config\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.308922 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwphc\" (UniqueName: \"kubernetes.io/projected/5052ce64-e1d7-429f-9256-f004ef04bc5a-kube-api-access-hwphc\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.308988 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-config\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309049 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309074 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309146 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309209 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6stp4\" (UniqueName: \"kubernetes.io/projected/a5724411-e0dd-4ab9-a898-0ecfbfe64385-kube-api-access-6stp4\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309342 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-config\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309394 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-ovndb-tls-certs\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309453 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-combined-ca-bundle\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.309995 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-svc\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.310576 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.312187 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.312874 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.313663 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-config\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.315792 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-combined-ca-bundle\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.318472 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-ovndb-tls-certs\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.325311 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-config\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.327003 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-httpd-config\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.340925 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6stp4\" (UniqueName: \"kubernetes.io/projected/a5724411-e0dd-4ab9-a898-0ecfbfe64385-kube-api-access-6stp4\") pod \"neutron-66cb4fcc98-c5psq\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.342689 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwphc\" (UniqueName: \"kubernetes.io/projected/5052ce64-e1d7-429f-9256-f004ef04bc5a-kube-api-access-hwphc\") pod \"dnsmasq-dns-55f844cf75-9g6ss\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.456699 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.483846 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e8ba59d-d691-47df-848b-f208dfe5fffb" path="/var/lib/kubelet/pods/2e8ba59d-d691-47df-848b-f208dfe5fffb/volumes" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.484661 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3069acac-b31f-4fa1-b902-023722aa791a" path="/var/lib/kubelet/pods/3069acac-b31f-4fa1-b902-023722aa791a/volumes" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.486063 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.488065 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40d40c68-19b2-4d91-834b-d54ffde9806d" path="/var/lib/kubelet/pods/40d40c68-19b2-4d91-834b-d54ffde9806d/volumes" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.489798 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5919e153-3626-4764-8110-ae849b21ff17" path="/var/lib/kubelet/pods/5919e153-3626-4764-8110-ae849b21ff17/volumes" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.505271 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5df99f8898-76z9s" event={"ID":"3407e6aa-fc06-4a2e-bd97-b8540ae12167","Type":"ContainerStarted","Data":"9585517a3e9e205d346954ccacb0c4ce059338dd017dc51cc71ce46f45fc2f15"} Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.520687 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-28kmg" event={"ID":"af15cc1d-88dc-44f9-b43c-9570dfc4fd65","Type":"ContainerStarted","Data":"7c06500225e2b9e067576af209242ded3b8902bc6166876744edd7adb866dcad"} Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.536940 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerStarted","Data":"e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee"} Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.547460 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-28kmg" podStartSLOduration=2.886424993 podStartE2EDuration="50.547442788s" podCreationTimestamp="2025-10-02 07:34:01 +0000 UTC" firstStartedPulling="2025-10-02 07:34:03.377341236 +0000 UTC m=+1034.716989641" lastFinishedPulling="2025-10-02 07:34:51.038359031 +0000 UTC m=+1082.378007436" observedRunningTime="2025-10-02 07:34:51.542708225 +0000 UTC m=+1082.882356630" watchObservedRunningTime="2025-10-02 07:34:51.547442788 +0000 UTC m=+1082.887091193" Oct 02 07:34:51 crc kubenswrapper[4829]: E1002 07:34:51.569683 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-xhn5v" podUID="ea5c0100-6e79-43fb-9761-62c7e758e891" Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.658536 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7dcfb8bd44-ws7cq"] Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.731720 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.775054 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-g5dzf"] Oct 02 07:34:51 crc kubenswrapper[4829]: W1002 07:34:51.796863 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode43b663f_cfca_459a_b513_ed16fdecd2f6.slice/crio-e297f06ff2fc4a2964a1999652851793cceceb21e49f46fe4e17d6282ad76c83 WatchSource:0}: Error finding container e297f06ff2fc4a2964a1999652851793cceceb21e49f46fe4e17d6282ad76c83: Status 404 returned error can't find the container with id e297f06ff2fc4a2964a1999652851793cceceb21e49f46fe4e17d6282ad76c83 Oct 02 07:34:51 crc kubenswrapper[4829]: I1002 07:34:51.909759 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-kj4jn" podUID="5919e153-3626-4764-8110-ae849b21ff17" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Oct 02 07:34:52 crc kubenswrapper[4829]: W1002 07:34:52.084840 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5052ce64_e1d7_429f_9256_f004ef04bc5a.slice/crio-9d366d589dfc4d9c501814df7cfe8b5d5f033da793dd12d0b72f056054caef21 WatchSource:0}: Error finding container 9d366d589dfc4d9c501814df7cfe8b5d5f033da793dd12d0b72f056054caef21: Status 404 returned error can't find the container with id 9d366d589dfc4d9c501814df7cfe8b5d5f033da793dd12d0b72f056054caef21 Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.133022 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-9g6ss"] Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.256017 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66cb4fcc98-c5psq"] Oct 02 07:34:52 crc kubenswrapper[4829]: W1002 07:34:52.264326 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5724411_e0dd_4ab9_a898_0ecfbfe64385.slice/crio-465c00156d0bf7ba97844390214dcf7b02c0353e500693b5086803ddd3958989 WatchSource:0}: Error finding container 465c00156d0bf7ba97844390214dcf7b02c0353e500693b5086803ddd3958989: Status 404 returned error can't find the container with id 465c00156d0bf7ba97844390214dcf7b02c0353e500693b5086803ddd3958989 Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.553126 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63630a30-0c56-4bdb-b259-984b81733bcc","Type":"ContainerStarted","Data":"951bbf7fcadc6852f1e4a91df2a509db3739c51d2b1ebfec874802ff3786f3dc"} Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.555429 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dcfb8bd44-ws7cq" event={"ID":"3a176a27-612f-499a-82c4-645d9cdd0b8b","Type":"ContainerStarted","Data":"b67312416317da230b283b4ee1b885cccef967d894ccafcc41f72dc87b5d9b0c"} Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.557261 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" event={"ID":"5052ce64-e1d7-429f-9256-f004ef04bc5a","Type":"ContainerStarted","Data":"9d366d589dfc4d9c501814df7cfe8b5d5f033da793dd12d0b72f056054caef21"} Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.558577 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cb4fcc98-c5psq" event={"ID":"a5724411-e0dd-4ab9-a898-0ecfbfe64385","Type":"ContainerStarted","Data":"465c00156d0bf7ba97844390214dcf7b02c0353e500693b5086803ddd3958989"} Oct 02 07:34:52 crc kubenswrapper[4829]: I1002 07:34:52.560124 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g5dzf" event={"ID":"e43b663f-cfca-459a-b513-ed16fdecd2f6","Type":"ContainerStarted","Data":"e297f06ff2fc4a2964a1999652851793cceceb21e49f46fe4e17d6282ad76c83"} Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.299003 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:34:53 crc kubenswrapper[4829]: W1002 07:34:53.308842 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8104fa43_766f_4747_bfde_8efb050e4629.slice/crio-ade4731835022ced7e89b38c9645bcdbba6daa3cc3bd28b5fc1e545a473f0874 WatchSource:0}: Error finding container ade4731835022ced7e89b38c9645bcdbba6daa3cc3bd28b5fc1e545a473f0874: Status 404 returned error can't find the container with id ade4731835022ced7e89b38c9645bcdbba6daa3cc3bd28b5fc1e545a473f0874 Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.575187 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8104fa43-766f-4747-bfde-8efb050e4629","Type":"ContainerStarted","Data":"ade4731835022ced7e89b38c9645bcdbba6daa3cc3bd28b5fc1e545a473f0874"} Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.770674 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-79b8d665c5-dgn5f"] Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.773055 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.782042 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.782272 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.791310 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79b8d665c5-dgn5f"] Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863656 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8p5n\" (UniqueName: \"kubernetes.io/projected/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-kube-api-access-n8p5n\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863731 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-internal-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863773 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-config\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863796 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-combined-ca-bundle\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863815 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-ovndb-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863864 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-httpd-config\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.863893 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-public-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965336 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-httpd-config\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965393 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-public-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965458 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8p5n\" (UniqueName: \"kubernetes.io/projected/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-kube-api-access-n8p5n\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965492 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-internal-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965529 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-config\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965556 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-combined-ca-bundle\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.965573 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-ovndb-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.972161 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-httpd-config\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.972173 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-public-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.972972 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-config\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.973958 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-ovndb-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.974139 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-combined-ca-bundle\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.981762 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-internal-tls-certs\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:53 crc kubenswrapper[4829]: I1002 07:34:53.992723 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8p5n\" (UniqueName: \"kubernetes.io/projected/1a40a5fd-a82e-430c-a15d-e32f0a14f5f2-kube-api-access-n8p5n\") pod \"neutron-79b8d665c5-dgn5f\" (UID: \"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2\") " pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.116150 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.640355 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dcfb8bd44-ws7cq" event={"ID":"3a176a27-612f-499a-82c4-645d9cdd0b8b","Type":"ContainerStarted","Data":"9814c7fd97f5c331f361eb5c9bf44206b3cba730dc3c445f6f3835986ffb7bef"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.656977 4829 generic.go:334] "Generic (PLEG): container finished" podID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerID="97f6ef6f9592cc02d589748832968f9e172748f384d0646837adf5a09af6c799" exitCode=0 Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.657052 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" event={"ID":"5052ce64-e1d7-429f-9256-f004ef04bc5a","Type":"ContainerDied","Data":"97f6ef6f9592cc02d589748832968f9e172748f384d0646837adf5a09af6c799"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.671048 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cb4fcc98-c5psq" event={"ID":"a5724411-e0dd-4ab9-a898-0ecfbfe64385","Type":"ContainerStarted","Data":"d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.671090 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cb4fcc98-c5psq" event={"ID":"a5724411-e0dd-4ab9-a898-0ecfbfe64385","Type":"ContainerStarted","Data":"e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.672061 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.694571 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g5dzf" event={"ID":"e43b663f-cfca-459a-b513-ed16fdecd2f6","Type":"ContainerStarted","Data":"063a9351068734af6e6fa6977f5cd04774af913fd826b3020b82866f57834828"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.701243 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-79b8d665c5-dgn5f"] Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.702209 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5df99f8898-76z9s" event={"ID":"3407e6aa-fc06-4a2e-bd97-b8540ae12167","Type":"ContainerStarted","Data":"e8d16d6f26ba5d8ce887d9b9822d73b091fd4bd09acd2d28cf289d33262adb96"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.707157 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66cb4fcc98-c5psq" podStartSLOduration=3.7070799 podStartE2EDuration="3.7070799s" podCreationTimestamp="2025-10-02 07:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:54.69623569 +0000 UTC m=+1086.035884095" watchObservedRunningTime="2025-10-02 07:34:54.7070799 +0000 UTC m=+1086.046728315" Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.717006 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8104fa43-766f-4747-bfde-8efb050e4629","Type":"ContainerStarted","Data":"5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.724885 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-g5dzf" podStartSLOduration=18.724859323 podStartE2EDuration="18.724859323s" podCreationTimestamp="2025-10-02 07:34:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:54.715623806 +0000 UTC m=+1086.055272211" watchObservedRunningTime="2025-10-02 07:34:54.724859323 +0000 UTC m=+1086.064507728" Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.733210 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63630a30-0c56-4bdb-b259-984b81733bcc","Type":"ContainerStarted","Data":"41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b"} Oct 02 07:34:54 crc kubenswrapper[4829]: I1002 07:34:54.766346 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5df99f8898-76z9s" podStartSLOduration=23.228523268 podStartE2EDuration="23.766328132s" podCreationTimestamp="2025-10-02 07:34:31 +0000 UTC" firstStartedPulling="2025-10-02 07:34:50.98812991 +0000 UTC m=+1082.327778325" lastFinishedPulling="2025-10-02 07:34:51.525934784 +0000 UTC m=+1082.865583189" observedRunningTime="2025-10-02 07:34:54.744416405 +0000 UTC m=+1086.084064810" watchObservedRunningTime="2025-10-02 07:34:54.766328132 +0000 UTC m=+1086.105976537" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.332623 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.332910 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.759617 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8tjd9" event={"ID":"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4","Type":"ContainerStarted","Data":"9a91941185e1b0c50fbbe0c382ab6f2c4f9da63274f907f001279591b7beb02c"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.770708 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8104fa43-766f-4747-bfde-8efb050e4629","Type":"ContainerStarted","Data":"54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.777172 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b8d665c5-dgn5f" event={"ID":"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2","Type":"ContainerStarted","Data":"b939abc582c9432a28d00268555a2d1769d5ede84aa655a1bd02575157589216"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.777215 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b8d665c5-dgn5f" event={"ID":"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2","Type":"ContainerStarted","Data":"85d6a3a00ed15ebf67b1d9b48b7e4533bd524566e956413b9891abade540a47e"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.779584 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-8tjd9" podStartSLOduration=3.98744431 podStartE2EDuration="33.779570356s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="2025-10-02 07:34:24.823835774 +0000 UTC m=+1056.163484179" lastFinishedPulling="2025-10-02 07:34:54.61596183 +0000 UTC m=+1085.955610225" observedRunningTime="2025-10-02 07:34:55.772968953 +0000 UTC m=+1087.112617358" watchObservedRunningTime="2025-10-02 07:34:55.779570356 +0000 UTC m=+1087.119218751" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.782945 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" event={"ID":"5052ce64-e1d7-429f-9256-f004ef04bc5a","Type":"ContainerStarted","Data":"0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.783625 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.785312 4829 generic.go:334] "Generic (PLEG): container finished" podID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" containerID="7c06500225e2b9e067576af209242ded3b8902bc6166876744edd7adb866dcad" exitCode=0 Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.785370 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-28kmg" event={"ID":"af15cc1d-88dc-44f9-b43c-9570dfc4fd65","Type":"ContainerDied","Data":"7c06500225e2b9e067576af209242ded3b8902bc6166876744edd7adb866dcad"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.799862 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=22.799838839 podStartE2EDuration="22.799838839s" podCreationTimestamp="2025-10-02 07:34:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:55.789802216 +0000 UTC m=+1087.129450621" watchObservedRunningTime="2025-10-02 07:34:55.799838839 +0000 UTC m=+1087.139487244" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.807409 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5df99f8898-76z9s" event={"ID":"3407e6aa-fc06-4a2e-bd97-b8540ae12167","Type":"ContainerStarted","Data":"c209e461815096a9937d2c4f915437bd97f194aa69ae75fa0d24d4f794bf6629"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.819940 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63630a30-0c56-4bdb-b259-984b81733bcc","Type":"ContainerStarted","Data":"6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.823502 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" podStartSLOduration=4.823480373 podStartE2EDuration="4.823480373s" podCreationTimestamp="2025-10-02 07:34:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:55.819621128 +0000 UTC m=+1087.159269533" watchObservedRunningTime="2025-10-02 07:34:55.823480373 +0000 UTC m=+1087.163128778" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.825292 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dcfb8bd44-ws7cq" event={"ID":"3a176a27-612f-499a-82c4-645d9cdd0b8b","Type":"ContainerStarted","Data":"dae3209bf1f7bb8956a52297b8d81605ad2f3b0ca6b90bc870590478cc3ac44f"} Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.861145 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=22.861127598 podStartE2EDuration="22.861127598s" podCreationTimestamp="2025-10-02 07:34:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:55.856444706 +0000 UTC m=+1087.196093111" watchObservedRunningTime="2025-10-02 07:34:55.861127598 +0000 UTC m=+1087.200776003" Oct 02 07:34:55 crc kubenswrapper[4829]: I1002 07:34:55.880101 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7dcfb8bd44-ws7cq" podStartSLOduration=24.880080089 podStartE2EDuration="24.880080089s" podCreationTimestamp="2025-10-02 07:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:55.875715519 +0000 UTC m=+1087.215363924" watchObservedRunningTime="2025-10-02 07:34:55.880080089 +0000 UTC m=+1087.219728494" Oct 02 07:34:56 crc kubenswrapper[4829]: I1002 07:34:56.841170 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-79b8d665c5-dgn5f" event={"ID":"1a40a5fd-a82e-430c-a15d-e32f0a14f5f2","Type":"ContainerStarted","Data":"5e2ca261b15a5ba3516f1ed1eaa6f66c2fb5a8a1475551d29d1b8d18aa5de65f"} Oct 02 07:34:56 crc kubenswrapper[4829]: I1002 07:34:56.841599 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:34:56 crc kubenswrapper[4829]: I1002 07:34:56.844114 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerStarted","Data":"7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7"} Oct 02 07:34:56 crc kubenswrapper[4829]: I1002 07:34:56.880634 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-79b8d665c5-dgn5f" podStartSLOduration=3.880617104 podStartE2EDuration="3.880617104s" podCreationTimestamp="2025-10-02 07:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:34:56.878717852 +0000 UTC m=+1088.218366257" watchObservedRunningTime="2025-10-02 07:34:56.880617104 +0000 UTC m=+1088.220265509" Oct 02 07:34:57 crc kubenswrapper[4829]: E1002 07:34:57.220803 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8adee8c4_a6f1_42cb_ae32_a7cc57b911b4.slice/crio-conmon-9a91941185e1b0c50fbbe0c382ab6f2c4f9da63274f907f001279591b7beb02c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.287717 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.347368 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-db-sync-config-data\") pod \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.347478 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-combined-ca-bundle\") pod \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.347512 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c42qk\" (UniqueName: \"kubernetes.io/projected/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-kube-api-access-c42qk\") pod \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.347564 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-config-data\") pod \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\" (UID: \"af15cc1d-88dc-44f9-b43c-9570dfc4fd65\") " Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.368881 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "af15cc1d-88dc-44f9-b43c-9570dfc4fd65" (UID: "af15cc1d-88dc-44f9-b43c-9570dfc4fd65"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.372380 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-kube-api-access-c42qk" (OuterVolumeSpecName: "kube-api-access-c42qk") pod "af15cc1d-88dc-44f9-b43c-9570dfc4fd65" (UID: "af15cc1d-88dc-44f9-b43c-9570dfc4fd65"). InnerVolumeSpecName "kube-api-access-c42qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.407357 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-config-data" (OuterVolumeSpecName: "config-data") pod "af15cc1d-88dc-44f9-b43c-9570dfc4fd65" (UID: "af15cc1d-88dc-44f9-b43c-9570dfc4fd65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.421058 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af15cc1d-88dc-44f9-b43c-9570dfc4fd65" (UID: "af15cc1d-88dc-44f9-b43c-9570dfc4fd65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.449003 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.449038 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c42qk\" (UniqueName: \"kubernetes.io/projected/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-kube-api-access-c42qk\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.449049 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.449057 4829 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af15cc1d-88dc-44f9-b43c-9570dfc4fd65-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.879609 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-28kmg" event={"ID":"af15cc1d-88dc-44f9-b43c-9570dfc4fd65","Type":"ContainerDied","Data":"7dc4e2252ac561c9d5f40fcc6ef8312500f0292e14108bf52f58ddd3cfa5d3cc"} Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.879667 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dc4e2252ac561c9d5f40fcc6ef8312500f0292e14108bf52f58ddd3cfa5d3cc" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.879741 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-28kmg" Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.885879 4829 generic.go:334] "Generic (PLEG): container finished" podID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" containerID="9a91941185e1b0c50fbbe0c382ab6f2c4f9da63274f907f001279591b7beb02c" exitCode=0 Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.886163 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8tjd9" event={"ID":"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4","Type":"ContainerDied","Data":"9a91941185e1b0c50fbbe0c382ab6f2c4f9da63274f907f001279591b7beb02c"} Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.888805 4829 generic.go:334] "Generic (PLEG): container finished" podID="e43b663f-cfca-459a-b513-ed16fdecd2f6" containerID="063a9351068734af6e6fa6977f5cd04774af913fd826b3020b82866f57834828" exitCode=0 Oct 02 07:34:57 crc kubenswrapper[4829]: I1002 07:34:57.889984 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g5dzf" event={"ID":"e43b663f-cfca-459a-b513-ed16fdecd2f6","Type":"ContainerDied","Data":"063a9351068734af6e6fa6977f5cd04774af913fd826b3020b82866f57834828"} Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.100708 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:34:58 crc kubenswrapper[4829]: E1002 07:34:58.101130 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" containerName="watcher-db-sync" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.101143 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" containerName="watcher-db-sync" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.101375 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" containerName="watcher-db-sync" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.102348 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.116332 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-jksjr" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.117091 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.148360 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.160694 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-config-data\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.160761 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.160808 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.160829 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6rlh\" (UniqueName: \"kubernetes.io/projected/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-kube-api-access-g6rlh\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.160862 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-logs\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.222042 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.223973 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.228086 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.233942 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.258085 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.260107 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.262589 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-logs\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.262726 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-config-data\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.262763 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.262797 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.262815 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6rlh\" (UniqueName: \"kubernetes.io/projected/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-kube-api-access-g6rlh\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.266437 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.267152 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-logs\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.274346 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.279964 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.283159 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-config-data\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.291398 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.317850 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6rlh\" (UniqueName: \"kubernetes.io/projected/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-kube-api-access-g6rlh\") pod \"watcher-api-0\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367186 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6390f00-cfcd-4b1c-831d-fd355daac37e-logs\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367245 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsdfd\" (UniqueName: \"kubernetes.io/projected/c6390f00-cfcd-4b1c-831d-fd355daac37e-kube-api-access-rsdfd\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367299 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-config-data\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367346 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jgrh\" (UniqueName: \"kubernetes.io/projected/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-kube-api-access-8jgrh\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367369 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367391 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-logs\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367412 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367436 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.367455 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.436648 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.468752 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469041 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-logs\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469131 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469233 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469305 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469419 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6390f00-cfcd-4b1c-831d-fd355daac37e-logs\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469511 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsdfd\" (UniqueName: \"kubernetes.io/projected/c6390f00-cfcd-4b1c-831d-fd355daac37e-kube-api-access-rsdfd\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469624 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-config-data\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.469726 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jgrh\" (UniqueName: \"kubernetes.io/projected/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-kube-api-access-8jgrh\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.470614 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-logs\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.478580 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6390f00-cfcd-4b1c-831d-fd355daac37e-logs\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.482118 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.482995 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.484911 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6390f00-cfcd-4b1c-831d-fd355daac37e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.491466 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-config-data\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.498799 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.504740 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jgrh\" (UniqueName: \"kubernetes.io/projected/59d8c82e-9aca-41ca-9d4b-8f5ec2824132-kube-api-access-8jgrh\") pod \"watcher-applier-0\" (UID: \"59d8c82e-9aca-41ca-9d4b-8f5ec2824132\") " pod="openstack/watcher-applier-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.514764 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsdfd\" (UniqueName: \"kubernetes.io/projected/c6390f00-cfcd-4b1c-831d-fd355daac37e-kube-api-access-rsdfd\") pod \"watcher-decision-engine-0\" (UID: \"c6390f00-cfcd-4b1c-831d-fd355daac37e\") " pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.543881 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Oct 02 07:34:58 crc kubenswrapper[4829]: I1002 07:34:58.713299 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.555011 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.555672 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.725661 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-scripts\") pod \"e43b663f-cfca-459a-b513-ed16fdecd2f6\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.725902 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-credential-keys\") pod \"e43b663f-cfca-459a-b513-ed16fdecd2f6\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.725947 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpqsr\" (UniqueName: \"kubernetes.io/projected/e43b663f-cfca-459a-b513-ed16fdecd2f6-kube-api-access-tpqsr\") pod \"e43b663f-cfca-459a-b513-ed16fdecd2f6\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.725973 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r8xs\" (UniqueName: \"kubernetes.io/projected/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-kube-api-access-9r8xs\") pod \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726013 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-logs\") pod \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726051 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-combined-ca-bundle\") pod \"e43b663f-cfca-459a-b513-ed16fdecd2f6\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726084 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-scripts\") pod \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726100 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-config-data\") pod \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726124 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-fernet-keys\") pod \"e43b663f-cfca-459a-b513-ed16fdecd2f6\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726160 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-config-data\") pod \"e43b663f-cfca-459a-b513-ed16fdecd2f6\" (UID: \"e43b663f-cfca-459a-b513-ed16fdecd2f6\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726249 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-combined-ca-bundle\") pod \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\" (UID: \"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4\") " Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726439 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-logs" (OuterVolumeSpecName: "logs") pod "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" (UID: "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.726999 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.732965 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "e43b663f-cfca-459a-b513-ed16fdecd2f6" (UID: "e43b663f-cfca-459a-b513-ed16fdecd2f6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.733916 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-kube-api-access-9r8xs" (OuterVolumeSpecName: "kube-api-access-9r8xs") pod "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" (UID: "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4"). InnerVolumeSpecName "kube-api-access-9r8xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.748218 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43b663f-cfca-459a-b513-ed16fdecd2f6-kube-api-access-tpqsr" (OuterVolumeSpecName: "kube-api-access-tpqsr") pod "e43b663f-cfca-459a-b513-ed16fdecd2f6" (UID: "e43b663f-cfca-459a-b513-ed16fdecd2f6"). InnerVolumeSpecName "kube-api-access-tpqsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.765427 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-scripts" (OuterVolumeSpecName: "scripts") pod "e43b663f-cfca-459a-b513-ed16fdecd2f6" (UID: "e43b663f-cfca-459a-b513-ed16fdecd2f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.765566 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "e43b663f-cfca-459a-b513-ed16fdecd2f6" (UID: "e43b663f-cfca-459a-b513-ed16fdecd2f6"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.769380 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-scripts" (OuterVolumeSpecName: "scripts") pod "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" (UID: "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.771304 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-config-data" (OuterVolumeSpecName: "config-data") pod "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" (UID: "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.773492 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-config-data" (OuterVolumeSpecName: "config-data") pod "e43b663f-cfca-459a-b513-ed16fdecd2f6" (UID: "e43b663f-cfca-459a-b513-ed16fdecd2f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.797649 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" (UID: "8adee8c4-a6f1-42cb-ae32-a7cc57b911b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.802044 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e43b663f-cfca-459a-b513-ed16fdecd2f6" (UID: "e43b663f-cfca-459a-b513-ed16fdecd2f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828522 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828556 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828565 4829 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828574 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpqsr\" (UniqueName: \"kubernetes.io/projected/e43b663f-cfca-459a-b513-ed16fdecd2f6-kube-api-access-tpqsr\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828584 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r8xs\" (UniqueName: \"kubernetes.io/projected/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-kube-api-access-9r8xs\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828592 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828600 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828607 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828615 4829 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.828624 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43b663f-cfca-459a-b513-ed16fdecd2f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.873602 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.910976 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8tjd9" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.911028 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8tjd9" event={"ID":"8adee8c4-a6f1-42cb-ae32-a7cc57b911b4","Type":"ContainerDied","Data":"82777d608b8f25d0796f7087c28e6c8a0aa200b47bf06ab317465c159e89713e"} Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.911085 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82777d608b8f25d0796f7087c28e6c8a0aa200b47bf06ab317465c159e89713e" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.914665 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-g5dzf" event={"ID":"e43b663f-cfca-459a-b513-ed16fdecd2f6","Type":"ContainerDied","Data":"e297f06ff2fc4a2964a1999652851793cceceb21e49f46fe4e17d6282ad76c83"} Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.914711 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e297f06ff2fc4a2964a1999652851793cceceb21e49f46fe4e17d6282ad76c83" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.914799 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-g5dzf" Oct 02 07:34:59 crc kubenswrapper[4829]: I1002 07:34:59.952697 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.049563 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6b77fd5f6d-mmltx"] Oct 02 07:35:00 crc kubenswrapper[4829]: E1002 07:35:00.050840 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" containerName="placement-db-sync" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.051095 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" containerName="placement-db-sync" Oct 02 07:35:00 crc kubenswrapper[4829]: E1002 07:35:00.051274 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43b663f-cfca-459a-b513-ed16fdecd2f6" containerName="keystone-bootstrap" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.051367 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43b663f-cfca-459a-b513-ed16fdecd2f6" containerName="keystone-bootstrap" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.051730 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43b663f-cfca-459a-b513-ed16fdecd2f6" containerName="keystone-bootstrap" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.051822 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" containerName="placement-db-sync" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.053025 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.056692 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.056973 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.057146 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.057327 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.057516 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-qxrff" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.058056 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.062455 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-995ff44cb-bcb2z"] Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.064376 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.066489 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-zfpss" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.067348 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.067818 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.068261 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.068685 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.111336 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b77fd5f6d-mmltx"] Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.128166 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-995ff44cb-bcb2z"] Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134597 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-config-data\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134634 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-public-tls-certs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134665 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfk8b\" (UniqueName: \"kubernetes.io/projected/5d8b5f01-0374-4315-82d6-de60c41b6bad-kube-api-access-bfk8b\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134733 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-credential-keys\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134836 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-scripts\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134875 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-combined-ca-bundle\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.134950 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-internal-tls-certs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135033 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-internal-tls-certs\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135062 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-fernet-keys\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135204 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-public-tls-certs\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135321 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-scripts\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135463 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x659p\" (UniqueName: \"kubernetes.io/projected/f49ffa44-6b66-43b1-9201-afc0e8bb9311-kube-api-access-x659p\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135552 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-config-data\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135589 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f49ffa44-6b66-43b1-9201-afc0e8bb9311-logs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.135609 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-combined-ca-bundle\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.155517 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237129 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x659p\" (UniqueName: \"kubernetes.io/projected/f49ffa44-6b66-43b1-9201-afc0e8bb9311-kube-api-access-x659p\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237214 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-config-data\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237256 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f49ffa44-6b66-43b1-9201-afc0e8bb9311-logs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237274 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-combined-ca-bundle\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237301 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-config-data\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237319 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-public-tls-certs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237338 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfk8b\" (UniqueName: \"kubernetes.io/projected/5d8b5f01-0374-4315-82d6-de60c41b6bad-kube-api-access-bfk8b\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237356 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-credential-keys\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237386 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-scripts\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237408 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-combined-ca-bundle\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237445 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-internal-tls-certs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237485 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-internal-tls-certs\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237506 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-fernet-keys\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237548 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-public-tls-certs\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.237574 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-scripts\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.238766 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f49ffa44-6b66-43b1-9201-afc0e8bb9311-logs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.242493 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-scripts\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.244053 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-internal-tls-certs\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.244421 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-credential-keys\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.245671 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-public-tls-certs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.246426 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-config-data\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.248299 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-public-tls-certs\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.249266 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-internal-tls-certs\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.253088 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49ffa44-6b66-43b1-9201-afc0e8bb9311-combined-ca-bundle\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.253898 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-config-data\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.253975 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-scripts\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.254200 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-combined-ca-bundle\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.255958 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5d8b5f01-0374-4315-82d6-de60c41b6bad-fernet-keys\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.258790 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x659p\" (UniqueName: \"kubernetes.io/projected/f49ffa44-6b66-43b1-9201-afc0e8bb9311-kube-api-access-x659p\") pod \"placement-995ff44cb-bcb2z\" (UID: \"f49ffa44-6b66-43b1-9201-afc0e8bb9311\") " pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.258892 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfk8b\" (UniqueName: \"kubernetes.io/projected/5d8b5f01-0374-4315-82d6-de60c41b6bad-kube-api-access-bfk8b\") pod \"keystone-6b77fd5f6d-mmltx\" (UID: \"5d8b5f01-0374-4315-82d6-de60c41b6bad\") " pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.377233 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:00 crc kubenswrapper[4829]: I1002 07:35:00.398404 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.458392 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.508434 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.509551 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.541606 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5j67n"] Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.545634 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="dnsmasq-dns" containerID="cri-o://37562d498b8f6e6ff4c4bb710c27c7b0805b91e7634c26144c52d25284d34733" gracePeriod=10 Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.761721 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.761853 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.943366 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"af0e9bda-c4cf-45c6-84e7-82a87d51cde5","Type":"ContainerStarted","Data":"3a1fed828fe9d09d09cff52a6ec6352753826cdce10c30c39f9ba4e0d3c85481"} Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.945430 4829 generic.go:334] "Generic (PLEG): container finished" podID="f65a4787-32b1-4883-b6b3-895e412001fb" containerID="37562d498b8f6e6ff4c4bb710c27c7b0805b91e7634c26144c52d25284d34733" exitCode=0 Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.945495 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" event={"ID":"f65a4787-32b1-4883-b6b3-895e412001fb","Type":"ContainerDied","Data":"37562d498b8f6e6ff4c4bb710c27c7b0805b91e7634c26144c52d25284d34733"} Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.947612 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c6390f00-cfcd-4b1c-831d-fd355daac37e","Type":"ContainerStarted","Data":"4fc4ac2cd7da6909360226444b935fccf0c52d4bd1008f783e4979ddf6223575"} Oct 02 07:35:01 crc kubenswrapper[4829]: I1002 07:35:01.949888 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"59d8c82e-9aca-41ca-9d4b-8f5ec2824132","Type":"ContainerStarted","Data":"48d160f4b387773dbb854e0155ef3f96975da16847520bac8feabbf0150b08c9"} Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.709341 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.710381 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.710533 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.710545 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.719779 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.720531 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.720610 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.720623 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.766991 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.780456 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.812522 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:03 crc kubenswrapper[4829]: I1002 07:35:03.812623 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.065874 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.117432 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-sb\") pod \"f65a4787-32b1-4883-b6b3-895e412001fb\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.117470 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-config\") pod \"f65a4787-32b1-4883-b6b3-895e412001fb\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.117524 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-nb\") pod \"f65a4787-32b1-4883-b6b3-895e412001fb\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.117555 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-swift-storage-0\") pod \"f65a4787-32b1-4883-b6b3-895e412001fb\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.117670 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq257\" (UniqueName: \"kubernetes.io/projected/f65a4787-32b1-4883-b6b3-895e412001fb-kube-api-access-gq257\") pod \"f65a4787-32b1-4883-b6b3-895e412001fb\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.117691 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-svc\") pod \"f65a4787-32b1-4883-b6b3-895e412001fb\" (UID: \"f65a4787-32b1-4883-b6b3-895e412001fb\") " Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.123250 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65a4787-32b1-4883-b6b3-895e412001fb-kube-api-access-gq257" (OuterVolumeSpecName: "kube-api-access-gq257") pod "f65a4787-32b1-4883-b6b3-895e412001fb" (UID: "f65a4787-32b1-4883-b6b3-895e412001fb"). InnerVolumeSpecName "kube-api-access-gq257". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.221033 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq257\" (UniqueName: \"kubernetes.io/projected/f65a4787-32b1-4883-b6b3-895e412001fb-kube-api-access-gq257\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.277103 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f65a4787-32b1-4883-b6b3-895e412001fb" (UID: "f65a4787-32b1-4883-b6b3-895e412001fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.292347 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f65a4787-32b1-4883-b6b3-895e412001fb" (UID: "f65a4787-32b1-4883-b6b3-895e412001fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.313902 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-config" (OuterVolumeSpecName: "config") pod "f65a4787-32b1-4883-b6b3-895e412001fb" (UID: "f65a4787-32b1-4883-b6b3-895e412001fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.322484 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.322527 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.322539 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.323573 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f65a4787-32b1-4883-b6b3-895e412001fb" (UID: "f65a4787-32b1-4883-b6b3-895e412001fb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.341397 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f65a4787-32b1-4883-b6b3-895e412001fb" (UID: "f65a4787-32b1-4883-b6b3-895e412001fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.360554 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6b77fd5f6d-mmltx"] Oct 02 07:35:04 crc kubenswrapper[4829]: W1002 07:35:04.384793 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d8b5f01_0374_4315_82d6_de60c41b6bad.slice/crio-b80cf69b335041b3dfc2baadb70a87c62bb16b91fd2dbaef7cf56a857663600a WatchSource:0}: Error finding container b80cf69b335041b3dfc2baadb70a87c62bb16b91fd2dbaef7cf56a857663600a: Status 404 returned error can't find the container with id b80cf69b335041b3dfc2baadb70a87c62bb16b91fd2dbaef7cf56a857663600a Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.425119 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.425151 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f65a4787-32b1-4883-b6b3-895e412001fb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.476783 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-995ff44cb-bcb2z"] Oct 02 07:35:04 crc kubenswrapper[4829]: W1002 07:35:04.518520 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf49ffa44_6b66_43b1_9201_afc0e8bb9311.slice/crio-f92b3a6d903b5d893fecb25003d15133cb22994fd73058e95233c8f2ed791906 WatchSource:0}: Error finding container f92b3a6d903b5d893fecb25003d15133cb22994fd73058e95233c8f2ed791906: Status 404 returned error can't find the container with id f92b3a6d903b5d893fecb25003d15133cb22994fd73058e95233c8f2ed791906 Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.997333 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-995ff44cb-bcb2z" event={"ID":"f49ffa44-6b66-43b1-9201-afc0e8bb9311","Type":"ContainerStarted","Data":"f92b3a6d903b5d893fecb25003d15133cb22994fd73058e95233c8f2ed791906"} Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.999808 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"af0e9bda-c4cf-45c6-84e7-82a87d51cde5","Type":"ContainerStarted","Data":"8bbd6a179886ed8459a89115df219d0162490abd4948ef2008674d98ed97f85d"} Oct 02 07:35:04 crc kubenswrapper[4829]: I1002 07:35:04.999836 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"af0e9bda-c4cf-45c6-84e7-82a87d51cde5","Type":"ContainerStarted","Data":"6327b6c8eb27e51707c5f8a5c037c13e1541c5b2ae22d116ded02a2ee00d3a9b"} Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.000400 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.004699 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerStarted","Data":"36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090"} Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.009119 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" event={"ID":"f65a4787-32b1-4883-b6b3-895e412001fb","Type":"ContainerDied","Data":"8088103963fc9559d505cdc4dee1c867bccb596acf47b203d331ee3e7e6d1e67"} Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.009171 4829 scope.go:117] "RemoveContainer" containerID="37562d498b8f6e6ff4c4bb710c27c7b0805b91e7634c26144c52d25284d34733" Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.009345 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.019639 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b77fd5f6d-mmltx" event={"ID":"5d8b5f01-0374-4315-82d6-de60c41b6bad","Type":"ContainerStarted","Data":"65ad686ade7a15c821bd5a21ebb5a69dc9d202e31b678b81816da4913c5a4128"} Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.019686 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.019699 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6b77fd5f6d-mmltx" event={"ID":"5d8b5f01-0374-4315-82d6-de60c41b6bad","Type":"ContainerStarted","Data":"b80cf69b335041b3dfc2baadb70a87c62bb16b91fd2dbaef7cf56a857663600a"} Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.033259 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=7.033238265 podStartE2EDuration="7.033238265s" podCreationTimestamp="2025-10-02 07:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:05.02068973 +0000 UTC m=+1096.360338135" watchObservedRunningTime="2025-10-02 07:35:05.033238265 +0000 UTC m=+1096.372886670" Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.051693 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6b77fd5f6d-mmltx" podStartSLOduration=5.051672379 podStartE2EDuration="5.051672379s" podCreationTimestamp="2025-10-02 07:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:05.051259407 +0000 UTC m=+1096.390907822" watchObservedRunningTime="2025-10-02 07:35:05.051672379 +0000 UTC m=+1096.391320784" Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.088996 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5j67n"] Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.099585 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-5j67n"] Oct 02 07:35:05 crc kubenswrapper[4829]: I1002 07:35:05.474387 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" path="/var/lib/kubelet/pods/f65a4787-32b1-4883-b6b3-895e412001fb/volumes" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.026741 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xhn5v" event={"ID":"ea5c0100-6e79-43fb-9761-62c7e758e891","Type":"ContainerStarted","Data":"da10059a3c91c82cf054a9566d982591ce0f1d1cd01453a06405b8c20b831c66"} Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.050877 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xhn5v" podStartSLOduration=3.987838872 podStartE2EDuration="43.050862431s" podCreationTimestamp="2025-10-02 07:34:23 +0000 UTC" firstStartedPulling="2025-10-02 07:34:24.951187564 +0000 UTC m=+1056.290835969" lastFinishedPulling="2025-10-02 07:35:04.014211123 +0000 UTC m=+1095.353859528" observedRunningTime="2025-10-02 07:35:06.044186546 +0000 UTC m=+1097.383834951" watchObservedRunningTime="2025-10-02 07:35:06.050862431 +0000 UTC m=+1097.390510836" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.217884 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.218063 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.234896 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.333872 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.333981 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:35:06 crc kubenswrapper[4829]: I1002 07:35:06.471910 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 07:35:07 crc kubenswrapper[4829]: I1002 07:35:07.182077 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 07:35:07 crc kubenswrapper[4829]: I1002 07:35:07.638136 4829 scope.go:117] "RemoveContainer" containerID="6c98ae7a568dbee9d3183312d9fce81ec583f32185de85ba77df9e19bc072a29" Oct 02 07:35:08 crc kubenswrapper[4829]: I1002 07:35:08.047283 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-995ff44cb-bcb2z" event={"ID":"f49ffa44-6b66-43b1-9201-afc0e8bb9311","Type":"ContainerStarted","Data":"c35f57f70849a97350ea2de876116918be9a3333cf3f9e60e85db03019790e95"} Oct 02 07:35:08 crc kubenswrapper[4829]: I1002 07:35:08.438281 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 02 07:35:08 crc kubenswrapper[4829]: I1002 07:35:08.438328 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 07:35:08 crc kubenswrapper[4829]: I1002 07:35:08.442444 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 02 07:35:08 crc kubenswrapper[4829]: I1002 07:35:08.901298 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-5j67n" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.156:5353: i/o timeout" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.078999 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"59d8c82e-9aca-41ca-9d4b-8f5ec2824132","Type":"ContainerStarted","Data":"35cdebbd6a948414be48a6de3f979d4176a59e9f6aa32b9b726b3023d135f3b2"} Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.082771 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"c6390f00-cfcd-4b1c-831d-fd355daac37e","Type":"ContainerStarted","Data":"e1f15dfc78b74117d5af58a77f4032e8cc69f1cfe85de54abcfb8abc166d0e1b"} Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.084895 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-995ff44cb-bcb2z" event={"ID":"f49ffa44-6b66-43b1-9201-afc0e8bb9311","Type":"ContainerStarted","Data":"20de21c23223fb1f93f80eefd95ab1d4928b18c1bbca5315bbe5d56e143682fc"} Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.084976 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.084999 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.087207 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dtlnv" event={"ID":"1b3627e8-07df-4d85-a723-a763eb14c52b","Type":"ContainerStarted","Data":"c9e91ac655fc2f9771cef316a9d9401d7568fc0fa938f60f8c095dc36caec018"} Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.091115 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.100933 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=5.126776389 podStartE2EDuration="11.100918707s" podCreationTimestamp="2025-10-02 07:34:58 +0000 UTC" firstStartedPulling="2025-10-02 07:35:01.753432385 +0000 UTC m=+1093.093080790" lastFinishedPulling="2025-10-02 07:35:07.727574703 +0000 UTC m=+1099.067223108" observedRunningTime="2025-10-02 07:35:09.093108115 +0000 UTC m=+1100.432756530" watchObservedRunningTime="2025-10-02 07:35:09.100918707 +0000 UTC m=+1100.440567112" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.114450 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=5.124833856 podStartE2EDuration="11.114434013s" podCreationTimestamp="2025-10-02 07:34:58 +0000 UTC" firstStartedPulling="2025-10-02 07:35:01.743170294 +0000 UTC m=+1093.082818699" lastFinishedPulling="2025-10-02 07:35:07.732770451 +0000 UTC m=+1099.072418856" observedRunningTime="2025-10-02 07:35:09.112193621 +0000 UTC m=+1100.451842026" watchObservedRunningTime="2025-10-02 07:35:09.114434013 +0000 UTC m=+1100.454082418" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.132131 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dtlnv" podStartSLOduration=3.963448157 podStartE2EDuration="47.132115504s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="2025-10-02 07:34:24.560236639 +0000 UTC m=+1055.899885044" lastFinishedPulling="2025-10-02 07:35:07.728903986 +0000 UTC m=+1099.068552391" observedRunningTime="2025-10-02 07:35:09.124299831 +0000 UTC m=+1100.463948236" watchObservedRunningTime="2025-10-02 07:35:09.132115504 +0000 UTC m=+1100.471763909" Oct 02 07:35:09 crc kubenswrapper[4829]: I1002 07:35:09.157067 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-995ff44cb-bcb2z" podStartSLOduration=9.157052587999999 podStartE2EDuration="9.157052588s" podCreationTimestamp="2025-10-02 07:35:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:09.149697191 +0000 UTC m=+1100.489345616" watchObservedRunningTime="2025-10-02 07:35:09.157052588 +0000 UTC m=+1100.496700993" Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.131714 4829 generic.go:334] "Generic (PLEG): container finished" podID="1b3627e8-07df-4d85-a723-a763eb14c52b" containerID="c9e91ac655fc2f9771cef316a9d9401d7568fc0fa938f60f8c095dc36caec018" exitCode=0 Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.132047 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dtlnv" event={"ID":"1b3627e8-07df-4d85-a723-a763eb14c52b","Type":"ContainerDied","Data":"c9e91ac655fc2f9771cef316a9d9401d7568fc0fa938f60f8c095dc36caec018"} Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.143565 4829 generic.go:334] "Generic (PLEG): container finished" podID="ea5c0100-6e79-43fb-9761-62c7e758e891" containerID="da10059a3c91c82cf054a9566d982591ce0f1d1cd01453a06405b8c20b831c66" exitCode=0 Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.143604 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xhn5v" event={"ID":"ea5c0100-6e79-43fb-9761-62c7e758e891","Type":"ContainerDied","Data":"da10059a3c91c82cf054a9566d982591ce0f1d1cd01453a06405b8c20b831c66"} Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.508444 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7dcfb8bd44-ws7cq" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.762717 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5df99f8898-76z9s" podUID="3407e6aa-fc06-4a2e-bd97-b8540ae12167" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.161:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.161:8443: connect: connection refused" Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.883723 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.883945 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api-log" containerID="cri-o://6327b6c8eb27e51707c5f8a5c037c13e1541c5b2ae22d116ded02a2ee00d3a9b" gracePeriod=30 Oct 02 07:35:11 crc kubenswrapper[4829]: I1002 07:35:11.884025 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api" containerID="cri-o://8bbd6a179886ed8459a89115df219d0162490abd4948ef2008674d98ed97f85d" gracePeriod=30 Oct 02 07:35:12 crc kubenswrapper[4829]: I1002 07:35:12.158792 4829 generic.go:334] "Generic (PLEG): container finished" podID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerID="6327b6c8eb27e51707c5f8a5c037c13e1541c5b2ae22d116ded02a2ee00d3a9b" exitCode=143 Oct 02 07:35:12 crc kubenswrapper[4829]: I1002 07:35:12.158962 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"af0e9bda-c4cf-45c6-84e7-82a87d51cde5","Type":"ContainerDied","Data":"6327b6c8eb27e51707c5f8a5c037c13e1541c5b2ae22d116ded02a2ee00d3a9b"} Oct 02 07:35:13 crc kubenswrapper[4829]: I1002 07:35:13.713915 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.025385 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.168:9322/\": read tcp 10.217.0.2:48718->10.217.0.168:9322: read: connection reset by peer" Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.025411 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.168:9322/\": read tcp 10.217.0.2:48726->10.217.0.168:9322: read: connection reset by peer" Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.190657 4829 generic.go:334] "Generic (PLEG): container finished" podID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerID="8bbd6a179886ed8459a89115df219d0162490abd4948ef2008674d98ed97f85d" exitCode=0 Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.190713 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"af0e9bda-c4cf-45c6-84e7-82a87d51cde5","Type":"ContainerDied","Data":"8bbd6a179886ed8459a89115df219d0162490abd4948ef2008674d98ed97f85d"} Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.875850 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.883916 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.989791 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x57f7\" (UniqueName: \"kubernetes.io/projected/ea5c0100-6e79-43fb-9761-62c7e758e891-kube-api-access-x57f7\") pod \"ea5c0100-6e79-43fb-9761-62c7e758e891\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.989993 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea5c0100-6e79-43fb-9761-62c7e758e891-etc-machine-id\") pod \"ea5c0100-6e79-43fb-9761-62c7e758e891\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.990099 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-db-sync-config-data\") pod \"1b3627e8-07df-4d85-a723-a763eb14c52b\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.990165 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-db-sync-config-data\") pod \"ea5c0100-6e79-43fb-9761-62c7e758e891\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.990276 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-combined-ca-bundle\") pod \"ea5c0100-6e79-43fb-9761-62c7e758e891\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.990828 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-combined-ca-bundle\") pod \"1b3627e8-07df-4d85-a723-a763eb14c52b\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.990933 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-scripts\") pod \"ea5c0100-6e79-43fb-9761-62c7e758e891\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.991070 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkwhv\" (UniqueName: \"kubernetes.io/projected/1b3627e8-07df-4d85-a723-a763eb14c52b-kube-api-access-lkwhv\") pod \"1b3627e8-07df-4d85-a723-a763eb14c52b\" (UID: \"1b3627e8-07df-4d85-a723-a763eb14c52b\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.991162 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-config-data\") pod \"ea5c0100-6e79-43fb-9761-62c7e758e891\" (UID: \"ea5c0100-6e79-43fb-9761-62c7e758e891\") " Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.991322 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea5c0100-6e79-43fb-9761-62c7e758e891-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ea5c0100-6e79-43fb-9761-62c7e758e891" (UID: "ea5c0100-6e79-43fb-9761-62c7e758e891"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:35:15 crc kubenswrapper[4829]: I1002 07:35:15.991656 4829 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ea5c0100-6e79-43fb-9761-62c7e758e891-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.009628 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea5c0100-6e79-43fb-9761-62c7e758e891-kube-api-access-x57f7" (OuterVolumeSpecName: "kube-api-access-x57f7") pod "ea5c0100-6e79-43fb-9761-62c7e758e891" (UID: "ea5c0100-6e79-43fb-9761-62c7e758e891"). InnerVolumeSpecName "kube-api-access-x57f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.009735 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1b3627e8-07df-4d85-a723-a763eb14c52b" (UID: "1b3627e8-07df-4d85-a723-a763eb14c52b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.017381 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b3627e8-07df-4d85-a723-a763eb14c52b-kube-api-access-lkwhv" (OuterVolumeSpecName: "kube-api-access-lkwhv") pod "1b3627e8-07df-4d85-a723-a763eb14c52b" (UID: "1b3627e8-07df-4d85-a723-a763eb14c52b"). InnerVolumeSpecName "kube-api-access-lkwhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.057437 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-scripts" (OuterVolumeSpecName: "scripts") pod "ea5c0100-6e79-43fb-9761-62c7e758e891" (UID: "ea5c0100-6e79-43fb-9761-62c7e758e891"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.058154 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ea5c0100-6e79-43fb-9761-62c7e758e891" (UID: "ea5c0100-6e79-43fb-9761-62c7e758e891"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.103362 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.103395 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkwhv\" (UniqueName: \"kubernetes.io/projected/1b3627e8-07df-4d85-a723-a763eb14c52b-kube-api-access-lkwhv\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.103405 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x57f7\" (UniqueName: \"kubernetes.io/projected/ea5c0100-6e79-43fb-9761-62c7e758e891-kube-api-access-x57f7\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.103414 4829 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.103423 4829 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.134660 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-config-data" (OuterVolumeSpecName: "config-data") pod "ea5c0100-6e79-43fb-9761-62c7e758e891" (UID: "ea5c0100-6e79-43fb-9761-62c7e758e891"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.148355 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea5c0100-6e79-43fb-9761-62c7e758e891" (UID: "ea5c0100-6e79-43fb-9761-62c7e758e891"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.148460 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b3627e8-07df-4d85-a723-a763eb14c52b" (UID: "1b3627e8-07df-4d85-a723-a763eb14c52b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.204552 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.204579 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea5c0100-6e79-43fb-9761-62c7e758e891-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.204589 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3627e8-07df-4d85-a723-a763eb14c52b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.238555 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xhn5v" event={"ID":"ea5c0100-6e79-43fb-9761-62c7e758e891","Type":"ContainerDied","Data":"6650b173eabc531f2dec4d9ded5eeb40e22f13dda30a8751d0be6b48b1ad8efb"} Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.238601 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6650b173eabc531f2dec4d9ded5eeb40e22f13dda30a8751d0be6b48b1ad8efb" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.238666 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xhn5v" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.247243 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dtlnv" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.247560 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dtlnv" event={"ID":"1b3627e8-07df-4d85-a723-a763eb14c52b","Type":"ContainerDied","Data":"45a2e7265e54e482cec6e946c0e1ff8372fdcdf6b661e4150c0c14a9edc9147c"} Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.247613 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45a2e7265e54e482cec6e946c0e1ff8372fdcdf6b661e4150c0c14a9edc9147c" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.329346 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.406959 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-custom-prometheus-ca\") pod \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.407040 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-combined-ca-bundle\") pod \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.407159 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6rlh\" (UniqueName: \"kubernetes.io/projected/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-kube-api-access-g6rlh\") pod \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.407248 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-logs\") pod \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.407280 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-config-data\") pod \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\" (UID: \"af0e9bda-c4cf-45c6-84e7-82a87d51cde5\") " Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.407860 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-logs" (OuterVolumeSpecName: "logs") pod "af0e9bda-c4cf-45c6-84e7-82a87d51cde5" (UID: "af0e9bda-c4cf-45c6-84e7-82a87d51cde5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.411655 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-kube-api-access-g6rlh" (OuterVolumeSpecName: "kube-api-access-g6rlh") pod "af0e9bda-c4cf-45c6-84e7-82a87d51cde5" (UID: "af0e9bda-c4cf-45c6-84e7-82a87d51cde5"). InnerVolumeSpecName "kube-api-access-g6rlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.432906 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "af0e9bda-c4cf-45c6-84e7-82a87d51cde5" (UID: "af0e9bda-c4cf-45c6-84e7-82a87d51cde5"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.448387 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af0e9bda-c4cf-45c6-84e7-82a87d51cde5" (UID: "af0e9bda-c4cf-45c6-84e7-82a87d51cde5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.472362 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-config-data" (OuterVolumeSpecName: "config-data") pod "af0e9bda-c4cf-45c6-84e7-82a87d51cde5" (UID: "af0e9bda-c4cf-45c6-84e7-82a87d51cde5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.510509 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.510546 4829 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.510562 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.510576 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6rlh\" (UniqueName: \"kubernetes.io/projected/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-kube-api-access-g6rlh\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:16 crc kubenswrapper[4829]: I1002 07:35:16.510587 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af0e9bda-c4cf-45c6-84e7-82a87d51cde5-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.158268 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6f9d8b6c5-ds24l"] Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.159030 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea5c0100-6e79-43fb-9761-62c7e758e891" containerName="cinder-db-sync" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159046 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea5c0100-6e79-43fb-9761-62c7e758e891" containerName="cinder-db-sync" Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.159065 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="init" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159073 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="init" Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.159096 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="dnsmasq-dns" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159106 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="dnsmasq-dns" Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.159133 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api-log" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159142 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api-log" Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.159154 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3627e8-07df-4d85-a723-a763eb14c52b" containerName="barbican-db-sync" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159162 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3627e8-07df-4d85-a723-a763eb14c52b" containerName="barbican-db-sync" Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.159185 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159193 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159434 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3627e8-07df-4d85-a723-a763eb14c52b" containerName="barbican-db-sync" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159453 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159478 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea5c0100-6e79-43fb-9761-62c7e758e891" containerName="cinder-db-sync" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159492 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65a4787-32b1-4883-b6b3-895e412001fb" containerName="dnsmasq-dns" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.159507 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" containerName="watcher-api-log" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.160751 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.165080 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.165356 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.165482 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-pv8ch" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.193899 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f9d8b6c5-ds24l"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.269389 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7cf66b5f98-d625p"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.270881 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.274514 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.290614 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7cf66b5f98-d625p"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.291048 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerStarted","Data":"51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557"} Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.291191 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-central-agent" containerID="cri-o://e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee" gracePeriod=30 Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.291436 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.291480 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="proxy-httpd" containerID="cri-o://51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557" gracePeriod=30 Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.291520 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="sg-core" containerID="cri-o://36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090" gracePeriod=30 Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.291553 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-notification-agent" containerID="cri-o://7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7" gracePeriod=30 Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.310540 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"af0e9bda-c4cf-45c6-84e7-82a87d51cde5","Type":"ContainerDied","Data":"3a1fed828fe9d09d09cff52a6ec6352753826cdce10c30c39f9ba4e0d3c85481"} Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.310611 4829 scope.go:117] "RemoveContainer" containerID="8bbd6a179886ed8459a89115df219d0162490abd4948ef2008674d98ed97f85d" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.310828 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.320743 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.322246 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.327154 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.327398 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4dmnc" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.331784 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.331895 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.335852 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.337209 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-config-data-custom\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.337248 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-combined-ca-bundle\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.337358 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54708879-c00e-4582-a37e-5becf429dc5d-logs\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.337483 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-config-data\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.337558 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stclw\" (UniqueName: \"kubernetes.io/projected/54708879-c00e-4582-a37e-5becf429dc5d-kube-api-access-stclw\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.369418 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.583462845 podStartE2EDuration="55.369395346s" podCreationTimestamp="2025-10-02 07:34:22 +0000 UTC" firstStartedPulling="2025-10-02 07:34:24.127407913 +0000 UTC m=+1055.467056318" lastFinishedPulling="2025-10-02 07:35:15.913340414 +0000 UTC m=+1107.252988819" observedRunningTime="2025-10-02 07:35:17.350530948 +0000 UTC m=+1108.690179363" watchObservedRunningTime="2025-10-02 07:35:17.369395346 +0000 UTC m=+1108.709043751" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.390396 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b895b5785-x9hfl"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.392164 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.397753 4829 scope.go:117] "RemoveContainer" containerID="6327b6c8eb27e51707c5f8a5c037c13e1541c5b2ae22d116ded02a2ee00d3a9b" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.411782 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-x9hfl"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.440871 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.440965 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e09c564a-17db-4723-a709-4fa0e897af52-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441026 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54708879-c00e-4582-a37e-5becf429dc5d-logs\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441074 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzhtx\" (UniqueName: \"kubernetes.io/projected/e09c564a-17db-4723-a709-4fa0e897af52-kube-api-access-qzhtx\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441105 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-combined-ca-bundle\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441174 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-config-data\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441242 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441266 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-scripts\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441289 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-config-data\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441310 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m75k\" (UniqueName: \"kubernetes.io/projected/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-kube-api-access-9m75k\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441338 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stclw\" (UniqueName: \"kubernetes.io/projected/54708879-c00e-4582-a37e-5becf429dc5d-kube-api-access-stclw\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441420 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441459 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-logs\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441555 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-config-data-custom\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441580 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-combined-ca-bundle\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.441623 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-config-data-custom\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.442292 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/54708879-c00e-4582-a37e-5becf429dc5d-logs\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.444304 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.450014 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-combined-ca-bundle\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.459035 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-config-data\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.461342 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54708879-c00e-4582-a37e-5becf429dc5d-config-data-custom\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.469801 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stclw\" (UniqueName: \"kubernetes.io/projected/54708879-c00e-4582-a37e-5becf429dc5d-kube-api-access-stclw\") pod \"barbican-worker-6f9d8b6c5-ds24l\" (UID: \"54708879-c00e-4582-a37e-5becf429dc5d\") " pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.476710 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f9d8b6c5-ds24l" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.509911 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.509950 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.517672 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.517772 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.521314 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.521635 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549322 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549379 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-scripts\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549416 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-config-data\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549445 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m75k\" (UniqueName: \"kubernetes.io/projected/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-kube-api-access-9m75k\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549495 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-config\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549519 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6kn5\" (UniqueName: \"kubernetes.io/projected/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-kube-api-access-w6kn5\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549547 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549583 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549608 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-logs\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549676 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549703 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-config-data-custom\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549742 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549763 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549794 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e09c564a-17db-4723-a709-4fa0e897af52-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549832 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzhtx\" (UniqueName: \"kubernetes.io/projected/e09c564a-17db-4723-a709-4fa0e897af52-kube-api-access-qzhtx\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549856 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-combined-ca-bundle\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.549956 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-svc\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.557915 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-logs\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.558525 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e09c564a-17db-4723-a709-4fa0e897af52-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.569005 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-combined-ca-bundle\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.570615 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.577350 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-config-data\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.581263 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.585473 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.586500 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-config-data-custom\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.590474 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.594215 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-scripts\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.649417 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-x9hfl"] Oct 02 07:35:17 crc kubenswrapper[4829]: E1002 07:35:17.653445 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-w6kn5 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-b895b5785-x9hfl" podUID="9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.660997 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m75k\" (UniqueName: \"kubernetes.io/projected/1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244-kube-api-access-9m75k\") pod \"barbican-keystone-listener-7cf66b5f98-d625p\" (UID: \"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244\") " pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.661706 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-svc\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.661843 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.661946 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab94865b-0ce8-4b10-bbab-7354e3603e1b-logs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.662198 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-config\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664067 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6kn5\" (UniqueName: \"kubernetes.io/projected/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-kube-api-access-w6kn5\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664245 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-config-data\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664349 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664440 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664575 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664696 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664767 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-public-tls-certs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664845 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spfw4\" (UniqueName: \"kubernetes.io/projected/ab94865b-0ce8-4b10-bbab-7354e3603e1b-kube-api-access-spfw4\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.664921 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.661889 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzhtx\" (UniqueName: \"kubernetes.io/projected/e09c564a-17db-4723-a709-4fa0e897af52-kube-api-access-qzhtx\") pod \"cinder-scheduler-0\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.666019 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-sb\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.666621 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-swift-storage-0\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.669992 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-nb\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.672763 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-config\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.685504 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.692130 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-svc\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.693485 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cmmvh"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.696921 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.701166 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6kn5\" (UniqueName: \"kubernetes.io/projected/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-kube-api-access-w6kn5\") pod \"dnsmasq-dns-b895b5785-x9hfl\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.776777 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.777051 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-public-tls-certs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.777070 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spfw4\" (UniqueName: \"kubernetes.io/projected/ab94865b-0ce8-4b10-bbab-7354e3603e1b-kube-api-access-spfw4\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.777149 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.777167 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab94865b-0ce8-4b10-bbab-7354e3603e1b-logs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.777256 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-config-data\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.777275 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.787209 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab94865b-0ce8-4b10-bbab-7354e3603e1b-logs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.833750 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.840851 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.844761 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-public-tls-certs\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.854538 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-config-data\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.859921 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/ab94865b-0ce8-4b10-bbab-7354e3603e1b-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.879643 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spfw4\" (UniqueName: \"kubernetes.io/projected/ab94865b-0ce8-4b10-bbab-7354e3603e1b-kube-api-access-spfw4\") pod \"watcher-api-0\" (UID: \"ab94865b-0ce8-4b10-bbab-7354e3603e1b\") " pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.880588 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.880630 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-config\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.880730 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.880758 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.880777 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.880798 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9hh\" (UniqueName: \"kubernetes.io/projected/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-kube-api-access-4x9hh\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.886178 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cmmvh"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.890780 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.939522 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7b44f6ccc4-59gl5"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.942255 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.944060 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.969298 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.971041 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.971745 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.974596 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.982446 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.982500 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.982522 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.982544 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9hh\" (UniqueName: \"kubernetes.io/projected/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-kube-api-access-4x9hh\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.982580 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.982608 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-config\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.983499 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-config\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.984005 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.984531 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.985013 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.985739 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:17 crc kubenswrapper[4829]: I1002 07:35:17.990302 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b44f6ccc4-59gl5"] Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.009601 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9hh\" (UniqueName: \"kubernetes.io/projected/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-kube-api-access-4x9hh\") pod \"dnsmasq-dns-5c9776ccc5-cmmvh\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.018818 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.085381 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkt5k\" (UniqueName: \"kubernetes.io/projected/de81169d-7ccd-4605-8fc3-ae24d838d0e1-kube-api-access-wkt5k\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.085639 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data-custom\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.085989 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-combined-ca-bundle\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086068 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086144 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data-custom\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086167 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086377 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086462 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqd6z\" (UniqueName: \"kubernetes.io/projected/f5eeddb3-db01-4916-94c8-a4914b521a0b-kube-api-access-jqd6z\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086506 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-scripts\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086527 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5eeddb3-db01-4916-94c8-a4914b521a0b-logs\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086605 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de81169d-7ccd-4605-8fc3-ae24d838d0e1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.086619 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de81169d-7ccd-4605-8fc3-ae24d838d0e1-logs\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188130 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188183 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqd6z\" (UniqueName: \"kubernetes.io/projected/f5eeddb3-db01-4916-94c8-a4914b521a0b-kube-api-access-jqd6z\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188203 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-scripts\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188263 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5eeddb3-db01-4916-94c8-a4914b521a0b-logs\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188292 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de81169d-7ccd-4605-8fc3-ae24d838d0e1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188306 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de81169d-7ccd-4605-8fc3-ae24d838d0e1-logs\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188347 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkt5k\" (UniqueName: \"kubernetes.io/projected/de81169d-7ccd-4605-8fc3-ae24d838d0e1-kube-api-access-wkt5k\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188423 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data-custom\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188438 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-combined-ca-bundle\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188458 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188482 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data-custom\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.188508 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.189624 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de81169d-7ccd-4605-8fc3-ae24d838d0e1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.190451 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5eeddb3-db01-4916-94c8-a4914b521a0b-logs\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.191066 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de81169d-7ccd-4605-8fc3-ae24d838d0e1-logs\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.193524 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-scripts\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.196099 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data-custom\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.196208 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.198447 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data-custom\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.207180 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-combined-ca-bundle\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.208295 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.218198 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkt5k\" (UniqueName: \"kubernetes.io/projected/de81169d-7ccd-4605-8fc3-ae24d838d0e1-kube-api-access-wkt5k\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.229697 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqd6z\" (UniqueName: \"kubernetes.io/projected/f5eeddb3-db01-4916-94c8-a4914b521a0b-kube-api-access-jqd6z\") pod \"barbican-api-7b44f6ccc4-59gl5\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.240424 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338395 4829 generic.go:334] "Generic (PLEG): container finished" podID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerID="51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557" exitCode=0 Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338426 4829 generic.go:334] "Generic (PLEG): container finished" podID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerID="36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090" exitCode=2 Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338435 4829 generic.go:334] "Generic (PLEG): container finished" podID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerID="e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee" exitCode=0 Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338488 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:18 crc kubenswrapper[4829]: E1002 07:35:18.338572 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode302e0a1_26ca_4e34_8edb_9c26df3f1812.slice/crio-51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338855 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerDied","Data":"51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557"} Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338886 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerDied","Data":"36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090"} Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.338896 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerDied","Data":"e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee"} Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.344482 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.363274 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.382149 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.389375 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.433819 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f9d8b6c5-ds24l"] Oct 02 07:35:18 crc kubenswrapper[4829]: W1002 07:35:18.452253 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54708879_c00e_4582_a37e_5becf429dc5d.slice/crio-61291b35d115969e32d17edd38fba9793e0b0e6bd217b443cc5b295c2734d03a WatchSource:0}: Error finding container 61291b35d115969e32d17edd38fba9793e0b0e6bd217b443cc5b295c2734d03a: Status 404 returned error can't find the container with id 61291b35d115969e32d17edd38fba9793e0b0e6bd217b443cc5b295c2734d03a Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.494178 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6kn5\" (UniqueName: \"kubernetes.io/projected/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-kube-api-access-w6kn5\") pod \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.494446 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-svc\") pod \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.494497 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-swift-storage-0\") pod \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.494568 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-sb\") pod \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.494593 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-config\") pod \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.494678 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-nb\") pod \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\" (UID: \"9a436aef-fbb8-45d2-bcb2-83e3cda21d0d\") " Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.497183 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" (UID: "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.499393 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" (UID: "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.499461 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" (UID: "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.500175 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" (UID: "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.501079 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-config" (OuterVolumeSpecName: "config") pod "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" (UID: "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.501601 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-kube-api-access-w6kn5" (OuterVolumeSpecName: "kube-api-access-w6kn5") pod "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" (UID: "9a436aef-fbb8-45d2-bcb2-83e3cda21d0d"). InnerVolumeSpecName "kube-api-access-w6kn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.545784 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.581050 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.597850 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.597923 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.597938 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.597949 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.597960 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.597972 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6kn5\" (UniqueName: \"kubernetes.io/projected/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d-kube-api-access-w6kn5\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.662882 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.717469 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Oct 02 07:35:18 crc kubenswrapper[4829]: W1002 07:35:18.751861 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode09c564a_17db_4723_a709_4fa0e897af52.slice/crio-638dab5bccff0224b3acbb7e49685978a3b4b6d4cab6c274cf1be43d35e8e5d9 WatchSource:0}: Error finding container 638dab5bccff0224b3acbb7e49685978a3b4b6d4cab6c274cf1be43d35e8e5d9: Status 404 returned error can't find the container with id 638dab5bccff0224b3acbb7e49685978a3b4b6d4cab6c274cf1be43d35e8e5d9 Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.758655 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.837356 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7cf66b5f98-d625p"] Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.846853 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Oct 02 07:35:18 crc kubenswrapper[4829]: I1002 07:35:18.952085 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cmmvh"] Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.114817 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:19 crc kubenswrapper[4829]: W1002 07:35:19.119514 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde81169d_7ccd_4605_8fc3_ae24d838d0e1.slice/crio-1cc38cb1d670fc1c941e4d840803eee9fd4e0f6c89882085ab0b79c9cf0b3bc6 WatchSource:0}: Error finding container 1cc38cb1d670fc1c941e4d840803eee9fd4e0f6c89882085ab0b79c9cf0b3bc6: Status 404 returned error can't find the container with id 1cc38cb1d670fc1c941e4d840803eee9fd4e0f6c89882085ab0b79c9cf0b3bc6 Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.123297 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b44f6ccc4-59gl5"] Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.368118 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f9d8b6c5-ds24l" event={"ID":"54708879-c00e-4582-a37e-5becf429dc5d","Type":"ContainerStarted","Data":"61291b35d115969e32d17edd38fba9793e0b0e6bd217b443cc5b295c2734d03a"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.369472 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e09c564a-17db-4723-a709-4fa0e897af52","Type":"ContainerStarted","Data":"638dab5bccff0224b3acbb7e49685978a3b4b6d4cab6c274cf1be43d35e8e5d9"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.371866 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"de81169d-7ccd-4605-8fc3-ae24d838d0e1","Type":"ContainerStarted","Data":"1cc38cb1d670fc1c941e4d840803eee9fd4e0f6c89882085ab0b79c9cf0b3bc6"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.377678 4829 generic.go:334] "Generic (PLEG): container finished" podID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerID="2040e619f09be56dae02ac8b949b77cfe20706963aac53700378978acbfe1f27" exitCode=0 Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.377744 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" event={"ID":"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8","Type":"ContainerDied","Data":"2040e619f09be56dae02ac8b949b77cfe20706963aac53700378978acbfe1f27"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.377768 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" event={"ID":"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8","Type":"ContainerStarted","Data":"c0dceb5d376687ff56e5fd5a20170b4e60bd7c4971b247c51f8bed97ed1ff692"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.383303 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" event={"ID":"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244","Type":"ContainerStarted","Data":"43850ef22c69144e261d055281d9e10fe6b19f13b0e3fb2f14170a6908e234b2"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.388563 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b44f6ccc4-59gl5" event={"ID":"f5eeddb3-db01-4916-94c8-a4914b521a0b","Type":"ContainerStarted","Data":"efc543cb030f8325744e88a64a8e4b333ad1f1408cc53b2a8ebf5a5c4b899919"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.403576 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"ab94865b-0ce8-4b10-bbab-7354e3603e1b","Type":"ContainerStarted","Data":"9a7a130191f1b031861978f510cf85947ccd0ab075d0f7f11fcdc60708e4a5c9"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.403607 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b895b5785-x9hfl" Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.403637 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"ab94865b-0ce8-4b10-bbab-7354e3603e1b","Type":"ContainerStarted","Data":"8dda86b1f8c2f7d0574f021420997f213c4b1378f04d44a568e17b9f2ab72be4"} Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.405299 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.518208 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0e9bda-c4cf-45c6-84e7-82a87d51cde5" path="/var/lib/kubelet/pods/af0e9bda-c4cf-45c6-84e7-82a87d51cde5/volumes" Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.518784 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-x9hfl"] Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.518806 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b895b5785-x9hfl"] Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.518870 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Oct 02 07:35:19 crc kubenswrapper[4829]: I1002 07:35:19.533653 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.086518 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242094 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-combined-ca-bundle\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242198 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-scripts\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242323 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-config-data\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242344 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-run-httpd\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242364 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8jtd\" (UniqueName: \"kubernetes.io/projected/e302e0a1-26ca-4e34-8edb-9c26df3f1812-kube-api-access-d8jtd\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242382 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-sg-core-conf-yaml\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.242412 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-log-httpd\") pod \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\" (UID: \"e302e0a1-26ca-4e34-8edb-9c26df3f1812\") " Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.243594 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.243892 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.277808 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-scripts" (OuterVolumeSpecName: "scripts") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.346785 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.346816 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.346827 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e302e0a1-26ca-4e34-8edb-9c26df3f1812-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.349552 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e302e0a1-26ca-4e34-8edb-9c26df3f1812-kube-api-access-d8jtd" (OuterVolumeSpecName: "kube-api-access-d8jtd") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "kube-api-access-d8jtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.390864 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.431912 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.435209 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.444294 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b44f6ccc4-59gl5" event={"ID":"f5eeddb3-db01-4916-94c8-a4914b521a0b","Type":"ContainerStarted","Data":"7d41ca2ed50d784ce82c011f41dac6aee40ce5935d199c629a92b077203e99f4"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.444349 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b44f6ccc4-59gl5" event={"ID":"f5eeddb3-db01-4916-94c8-a4914b521a0b","Type":"ContainerStarted","Data":"f407cf8ffe266d1518e3a446c838a1c806736bf5fa0f3aba4b5c0165ec7741b2"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.445954 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.445998 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.449280 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.449309 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8jtd\" (UniqueName: \"kubernetes.io/projected/e302e0a1-26ca-4e34-8edb-9c26df3f1812-kube-api-access-d8jtd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.449321 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.456806 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"ab94865b-0ce8-4b10-bbab-7354e3603e1b","Type":"ContainerStarted","Data":"1acd6c404dfd0b01a615958a4360fa8250b7784d74fe9498f4b0f9ebfd3d7b04"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.458773 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.481361 4829 generic.go:334] "Generic (PLEG): container finished" podID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerID="7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7" exitCode=0 Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.481447 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerDied","Data":"7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.481473 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e302e0a1-26ca-4e34-8edb-9c26df3f1812","Type":"ContainerDied","Data":"458fbbb2ee00e56a2eac486eff2fba6e6c15a408ee78771c85bffe90504ef928"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.481490 4829 scope.go:117] "RemoveContainer" containerID="51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.481629 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.483605 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7b44f6ccc4-59gl5" podStartSLOduration=3.4835914519999998 podStartE2EDuration="3.483591452s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:20.481668611 +0000 UTC m=+1111.821317016" watchObservedRunningTime="2025-10-02 07:35:20.483591452 +0000 UTC m=+1111.823239857" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.497573 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"de81169d-7ccd-4605-8fc3-ae24d838d0e1","Type":"ContainerStarted","Data":"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.502475 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" event={"ID":"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8","Type":"ContainerStarted","Data":"3c52ebe8e796e8bca72e4cb68184e3949bbee197dd0435c2787e4c7aa2243eb3"} Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.503652 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.563773 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=3.563746239 podStartE2EDuration="3.563746239s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:20.530666421 +0000 UTC m=+1111.870314826" watchObservedRunningTime="2025-10-02 07:35:20.563746239 +0000 UTC m=+1111.903394644" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.593413 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" podStartSLOduration=3.593395226 podStartE2EDuration="3.593395226s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:20.587465654 +0000 UTC m=+1111.927114059" watchObservedRunningTime="2025-10-02 07:35:20.593395226 +0000 UTC m=+1111.933043631" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.647056 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-config-data" (OuterVolumeSpecName: "config-data") pod "e302e0a1-26ca-4e34-8edb-9c26df3f1812" (UID: "e302e0a1-26ca-4e34-8edb-9c26df3f1812"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.658331 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e302e0a1-26ca-4e34-8edb-9c26df3f1812-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.832590 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.840436 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.860644 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4829]: E1002 07:35:20.861100 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-notification-agent" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861119 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-notification-agent" Oct 02 07:35:20 crc kubenswrapper[4829]: E1002 07:35:20.861143 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-central-agent" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861150 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-central-agent" Oct 02 07:35:20 crc kubenswrapper[4829]: E1002 07:35:20.861169 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="sg-core" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861175 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="sg-core" Oct 02 07:35:20 crc kubenswrapper[4829]: E1002 07:35:20.861201 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="proxy-httpd" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861207 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="proxy-httpd" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861447 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="proxy-httpd" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861469 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="sg-core" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861489 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-notification-agent" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.861510 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" containerName="ceilometer-central-agent" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.863218 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.870085 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.870685 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.873278 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.964964 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-log-httpd\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.965047 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws8hw\" (UniqueName: \"kubernetes.io/projected/f45b611a-62d2-4c65-8a5f-8557b05da53a-kube-api-access-ws8hw\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.965086 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.965238 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-run-httpd\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.965315 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-config-data\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.965498 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-scripts\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:20 crc kubenswrapper[4829]: I1002 07:35:20.965529 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.067471 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-log-httpd\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.067729 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws8hw\" (UniqueName: \"kubernetes.io/projected/f45b611a-62d2-4c65-8a5f-8557b05da53a-kube-api-access-ws8hw\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.067853 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.067970 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-run-httpd\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.068055 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-config-data\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.068160 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-scripts\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.068254 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.068872 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-log-httpd\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.069026 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-run-httpd\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.072514 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-config-data\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.073103 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.073906 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-scripts\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.085454 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.086823 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws8hw\" (UniqueName: \"kubernetes.io/projected/f45b611a-62d2-4c65-8a5f-8557b05da53a-kube-api-access-ws8hw\") pod \"ceilometer-0\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.194320 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.474953 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a436aef-fbb8-45d2-bcb2-83e3cda21d0d" path="/var/lib/kubelet/pods/9a436aef-fbb8-45d2-bcb2-83e3cda21d0d/volumes" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.475743 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e302e0a1-26ca-4e34-8edb-9c26df3f1812" path="/var/lib/kubelet/pods/e302e0a1-26ca-4e34-8edb-9c26df3f1812/volumes" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.508164 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.758469 4829 scope.go:117] "RemoveContainer" containerID="36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.814645 4829 scope.go:117] "RemoveContainer" containerID="7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7" Oct 02 07:35:21 crc kubenswrapper[4829]: I1002 07:35:21.977837 4829 scope.go:117] "RemoveContainer" containerID="e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.137763 4829 scope.go:117] "RemoveContainer" containerID="51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557" Oct 02 07:35:22 crc kubenswrapper[4829]: E1002 07:35:22.138546 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557\": container with ID starting with 51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557 not found: ID does not exist" containerID="51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.138574 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557"} err="failed to get container status \"51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557\": rpc error: code = NotFound desc = could not find container \"51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557\": container with ID starting with 51cb5e29346ac8316cdb6ccf85d9d56e3acb934f152af90edfc56737d92d8557 not found: ID does not exist" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.138595 4829 scope.go:117] "RemoveContainer" containerID="36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090" Oct 02 07:35:22 crc kubenswrapper[4829]: E1002 07:35:22.139288 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090\": container with ID starting with 36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090 not found: ID does not exist" containerID="36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.139314 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090"} err="failed to get container status \"36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090\": rpc error: code = NotFound desc = could not find container \"36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090\": container with ID starting with 36e548392e40f1e553890bd9845bc45cf83092e3a7fa2bfbd45a45510d521090 not found: ID does not exist" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.139327 4829 scope.go:117] "RemoveContainer" containerID="7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7" Oct 02 07:35:22 crc kubenswrapper[4829]: E1002 07:35:22.145935 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7\": container with ID starting with 7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7 not found: ID does not exist" containerID="7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.145964 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7"} err="failed to get container status \"7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7\": rpc error: code = NotFound desc = could not find container \"7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7\": container with ID starting with 7eea3c27033a9b987f600921360f7f89a6e88074774758659853ab6f43a398a7 not found: ID does not exist" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.145979 4829 scope.go:117] "RemoveContainer" containerID="e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee" Oct 02 07:35:22 crc kubenswrapper[4829]: E1002 07:35:22.146503 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee\": container with ID starting with e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee not found: ID does not exist" containerID="e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.146541 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee"} err="failed to get container status \"e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee\": rpc error: code = NotFound desc = could not find container \"e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee\": container with ID starting with e49d9c26cc9929be449954529eb2a37b07796c72eae3f622fbbcfc5cf5b6c4ee not found: ID does not exist" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.402154 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.546826 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f9d8b6c5-ds24l" event={"ID":"54708879-c00e-4582-a37e-5becf429dc5d","Type":"ContainerStarted","Data":"fc68aacaa6b03ec5dd55f9c8fa71c7bb23e3eb7b12f3c5620d5e150aaccbfefe"} Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.549745 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e09c564a-17db-4723-a709-4fa0e897af52","Type":"ContainerStarted","Data":"80372dd323dd942281116dc27256c3c31e050ed8081300fbed8a5c7eae8110aa"} Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.551048 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" event={"ID":"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244","Type":"ContainerStarted","Data":"4ee3bfdbc7ce2dc52cb485b7a3ab7690f778c8db020c696609ef32a2c12fedd6"} Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.552121 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerStarted","Data":"09869c638bfc07a67dc444c439b9ff6464cdc1ed8f57c08a97e6af6f6ae32439"} Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.553403 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.611469 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6f9d8b6c5-ds24l" podStartSLOduration=2.234265871 podStartE2EDuration="5.611447223s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="2025-10-02 07:35:18.458154677 +0000 UTC m=+1109.797803082" lastFinishedPulling="2025-10-02 07:35:21.835336029 +0000 UTC m=+1113.174984434" observedRunningTime="2025-10-02 07:35:22.585187605 +0000 UTC m=+1113.924836020" watchObservedRunningTime="2025-10-02 07:35:22.611447223 +0000 UTC m=+1113.951095628" Oct 02 07:35:22 crc kubenswrapper[4829]: I1002 07:35:22.972813 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.285837 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.564689 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e09c564a-17db-4723-a709-4fa0e897af52","Type":"ContainerStarted","Data":"35b374e52770c3e6f5f3af1948ff51327535721d20cab9ef4bb9e57b03499f43"} Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.569568 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"de81169d-7ccd-4605-8fc3-ae24d838d0e1","Type":"ContainerStarted","Data":"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc"} Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.569737 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api-log" containerID="cri-o://71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03" gracePeriod=30 Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.569999 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.570032 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api" containerID="cri-o://a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc" gracePeriod=30 Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.584536 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" event={"ID":"1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244","Type":"ContainerStarted","Data":"f44dd130ce7d64f6ef3f70d6e254fb2c7b041b575c1500d2d5f4b7b748f3a76b"} Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.587337 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerStarted","Data":"f82dba84ad8187641e0688eef29bb60204eb0d65f94df93c4f36e4d52426c8a6"} Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.594824 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f9d8b6c5-ds24l" event={"ID":"54708879-c00e-4582-a37e-5becf429dc5d","Type":"ContainerStarted","Data":"9435767e93186a4cfd6dac420dc2e336fee3cd041d1401ef52b793974d158b15"} Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.598505 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.401698454 podStartE2EDuration="6.598493721s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="2025-10-02 07:35:18.757667982 +0000 UTC m=+1110.097316387" lastFinishedPulling="2025-10-02 07:35:19.954463249 +0000 UTC m=+1111.294111654" observedRunningTime="2025-10-02 07:35:23.595942489 +0000 UTC m=+1114.935590894" watchObservedRunningTime="2025-10-02 07:35:23.598493721 +0000 UTC m=+1114.938142126" Oct 02 07:35:23 crc kubenswrapper[4829]: I1002 07:35:23.623568 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7cf66b5f98-d625p" podStartSLOduration=3.647807081 podStartE2EDuration="6.62354974s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="2025-10-02 07:35:18.859734265 +0000 UTC m=+1110.199382670" lastFinishedPulling="2025-10-02 07:35:21.835476924 +0000 UTC m=+1113.175125329" observedRunningTime="2025-10-02 07:35:23.619399826 +0000 UTC m=+1114.959048231" watchObservedRunningTime="2025-10-02 07:35:23.62354974 +0000 UTC m=+1114.963198145" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.121947 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.121929431 podStartE2EDuration="7.121929431s" podCreationTimestamp="2025-10-02 07:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:23.647064008 +0000 UTC m=+1114.986712413" watchObservedRunningTime="2025-10-02 07:35:24.121929431 +0000 UTC m=+1115.461577836" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.125309 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c7944ccd6-g5gh7"] Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.126818 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.131202 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7944ccd6-g5gh7"] Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.131805 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.132073 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.169061 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-79b8d665c5-dgn5f" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.228653 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66cb4fcc98-c5psq"] Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.228931 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66cb4fcc98-c5psq" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-api" containerID="cri-o://e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74" gracePeriod=30 Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.232144 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66cb4fcc98-c5psq" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-httpd" containerID="cri-o://d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85" gracePeriod=30 Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272502 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f201db-52b1-4a9f-a1f3-adb1bdb923da-logs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272580 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzlq8\" (UniqueName: \"kubernetes.io/projected/69f201db-52b1-4a9f-a1f3-adb1bdb923da-kube-api-access-dzlq8\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272625 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-combined-ca-bundle\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272650 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-internal-tls-certs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272719 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-public-tls-certs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272735 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-config-data-custom\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.272762 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-config-data\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376737 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzlq8\" (UniqueName: \"kubernetes.io/projected/69f201db-52b1-4a9f-a1f3-adb1bdb923da-kube-api-access-dzlq8\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376800 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-combined-ca-bundle\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376829 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-internal-tls-certs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376896 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-public-tls-certs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376915 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-config-data-custom\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376937 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-config-data\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.376994 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f201db-52b1-4a9f-a1f3-adb1bdb923da-logs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.377409 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69f201db-52b1-4a9f-a1f3-adb1bdb923da-logs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.387440 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-config-data\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.388103 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-config-data-custom\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.401421 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-public-tls-certs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.405885 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-combined-ca-bundle\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.411630 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzlq8\" (UniqueName: \"kubernetes.io/projected/69f201db-52b1-4a9f-a1f3-adb1bdb923da-kube-api-access-dzlq8\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.424616 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/69f201db-52b1-4a9f-a1f3-adb1bdb923da-internal-tls-certs\") pod \"barbican-api-6c7944ccd6-g5gh7\" (UID: \"69f201db-52b1-4a9f-a1f3-adb1bdb923da\") " pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.463799 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.520301 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.618784 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerStarted","Data":"fa5d98bb57809fcf6500388558873874ef292c270da6cfcdffc7088cba73cfc2"} Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630575 4829 generic.go:334] "Generic (PLEG): container finished" podID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerID="a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc" exitCode=0 Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630604 4829 generic.go:334] "Generic (PLEG): container finished" podID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerID="71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03" exitCode=143 Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630646 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"de81169d-7ccd-4605-8fc3-ae24d838d0e1","Type":"ContainerDied","Data":"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc"} Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630672 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"de81169d-7ccd-4605-8fc3-ae24d838d0e1","Type":"ContainerDied","Data":"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03"} Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630681 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"de81169d-7ccd-4605-8fc3-ae24d838d0e1","Type":"ContainerDied","Data":"1cc38cb1d670fc1c941e4d840803eee9fd4e0f6c89882085ab0b79c9cf0b3bc6"} Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630697 4829 scope.go:117] "RemoveContainer" containerID="a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.630828 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.638431 4829 generic.go:334] "Generic (PLEG): container finished" podID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerID="d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85" exitCode=0 Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.639063 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cb4fcc98-c5psq" event={"ID":"a5724411-e0dd-4ab9-a898-0ecfbfe64385","Type":"ContainerDied","Data":"d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85"} Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.663571 4829 scope.go:117] "RemoveContainer" containerID="71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.682576 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de81169d-7ccd-4605-8fc3-ae24d838d0e1-logs\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.682799 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkt5k\" (UniqueName: \"kubernetes.io/projected/de81169d-7ccd-4605-8fc3-ae24d838d0e1-kube-api-access-wkt5k\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.682850 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data-custom\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.682901 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de81169d-7ccd-4605-8fc3-ae24d838d0e1-etc-machine-id\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.682949 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.683007 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-combined-ca-bundle\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.683120 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-scripts\") pod \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\" (UID: \"de81169d-7ccd-4605-8fc3-ae24d838d0e1\") " Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.684457 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/de81169d-7ccd-4605-8fc3-ae24d838d0e1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.685433 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de81169d-7ccd-4605-8fc3-ae24d838d0e1-logs" (OuterVolumeSpecName: "logs") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.692348 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.692460 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de81169d-7ccd-4605-8fc3-ae24d838d0e1-kube-api-access-wkt5k" (OuterVolumeSpecName: "kube-api-access-wkt5k") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "kube-api-access-wkt5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.693411 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-scripts" (OuterVolumeSpecName: "scripts") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.695396 4829 scope.go:117] "RemoveContainer" containerID="a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc" Oct 02 07:35:24 crc kubenswrapper[4829]: E1002 07:35:24.695848 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc\": container with ID starting with a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc not found: ID does not exist" containerID="a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.695873 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc"} err="failed to get container status \"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc\": rpc error: code = NotFound desc = could not find container \"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc\": container with ID starting with a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc not found: ID does not exist" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.695894 4829 scope.go:117] "RemoveContainer" containerID="71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03" Oct 02 07:35:24 crc kubenswrapper[4829]: E1002 07:35:24.696766 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03\": container with ID starting with 71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03 not found: ID does not exist" containerID="71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.696793 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03"} err="failed to get container status \"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03\": rpc error: code = NotFound desc = could not find container \"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03\": container with ID starting with 71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03 not found: ID does not exist" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.696809 4829 scope.go:117] "RemoveContainer" containerID="a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.696997 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc"} err="failed to get container status \"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc\": rpc error: code = NotFound desc = could not find container \"a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc\": container with ID starting with a75a2ac5a4bb9285b9ad4d3ade46c894cf38fb0e127ca05c592e8a83c9aabfbc not found: ID does not exist" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.697010 4829 scope.go:117] "RemoveContainer" containerID="71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.697173 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03"} err="failed to get container status \"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03\": rpc error: code = NotFound desc = could not find container \"71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03\": container with ID starting with 71c86f1de7ac05b78992be88d32580bc554f86a587988faf489dc5d8cf1a0b03 not found: ID does not exist" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.754633 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.786107 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de81169d-7ccd-4605-8fc3-ae24d838d0e1-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.786136 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkt5k\" (UniqueName: \"kubernetes.io/projected/de81169d-7ccd-4605-8fc3-ae24d838d0e1-kube-api-access-wkt5k\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.786149 4829 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.786157 4829 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de81169d-7ccd-4605-8fc3-ae24d838d0e1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.786167 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.786176 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.800455 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data" (OuterVolumeSpecName: "config-data") pod "de81169d-7ccd-4605-8fc3-ae24d838d0e1" (UID: "de81169d-7ccd-4605-8fc3-ae24d838d0e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.880915 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.889684 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:35:24 crc kubenswrapper[4829]: I1002 07:35:24.895720 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de81169d-7ccd-4605-8fc3-ae24d838d0e1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.058766 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.088775 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.152379 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:25 crc kubenswrapper[4829]: E1002 07:35:25.152910 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api-log" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.152931 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api-log" Oct 02 07:35:25 crc kubenswrapper[4829]: E1002 07:35:25.152959 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.152968 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.153209 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api-log" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.153243 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" containerName="cinder-api" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.154388 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.159478 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.159670 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.159823 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.174186 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.204356 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c7944ccd6-g5gh7"] Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.328774 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.328825 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7zcp\" (UniqueName: \"kubernetes.io/projected/2bec1c3c-60b9-4776-93c0-3d51886b3552-kube-api-access-n7zcp\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332816 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332847 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-scripts\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332864 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bec1c3c-60b9-4776-93c0-3d51886b3552-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332890 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bec1c3c-60b9-4776-93c0-3d51886b3552-logs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332952 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332970 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-config-data\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.332997 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.333025 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-config-data-custom\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434141 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-scripts\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434186 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bec1c3c-60b9-4776-93c0-3d51886b3552-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434219 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bec1c3c-60b9-4776-93c0-3d51886b3552-logs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434311 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434332 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-config-data\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434361 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434398 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-config-data-custom\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434430 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7zcp\" (UniqueName: \"kubernetes.io/projected/2bec1c3c-60b9-4776-93c0-3d51886b3552-kube-api-access-n7zcp\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.434450 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.436617 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2bec1c3c-60b9-4776-93c0-3d51886b3552-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.437256 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2bec1c3c-60b9-4776-93c0-3d51886b3552-logs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.439720 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-scripts\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.440580 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-config-data\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.442789 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.443179 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.450405 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.450769 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bec1c3c-60b9-4776-93c0-3d51886b3552-config-data-custom\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.462466 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7zcp\" (UniqueName: \"kubernetes.io/projected/2bec1c3c-60b9-4776-93c0-3d51886b3552-kube-api-access-n7zcp\") pod \"cinder-api-0\" (UID: \"2bec1c3c-60b9-4776-93c0-3d51886b3552\") " pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.476778 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de81169d-7ccd-4605-8fc3-ae24d838d0e1" path="/var/lib/kubelet/pods/de81169d-7ccd-4605-8fc3-ae24d838d0e1/volumes" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.560533 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.676075 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7944ccd6-g5gh7" event={"ID":"69f201db-52b1-4a9f-a1f3-adb1bdb923da","Type":"ContainerStarted","Data":"2d1a356bcec5a83117c928eb1e5901029376815ab601d6b2b53c08823e2523b0"} Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.676121 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7944ccd6-g5gh7" event={"ID":"69f201db-52b1-4a9f-a1f3-adb1bdb923da","Type":"ContainerStarted","Data":"c5bc89a9ec00f8a77ac672be8e3b77ed289f1e134da0cee85e4d751de7a6e085"} Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.676132 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c7944ccd6-g5gh7" event={"ID":"69f201db-52b1-4a9f-a1f3-adb1bdb923da","Type":"ContainerStarted","Data":"22f082c17790692b8b7bb73003e59192010a46b93d32570d009f8d52a1c90a4e"} Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.676731 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.677106 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.708713 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c7944ccd6-g5gh7" podStartSLOduration=1.7086928719999999 podStartE2EDuration="1.708692872s" podCreationTimestamp="2025-10-02 07:35:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:25.696344112 +0000 UTC m=+1117.035992507" watchObservedRunningTime="2025-10-02 07:35:25.708692872 +0000 UTC m=+1117.048341277" Oct 02 07:35:25 crc kubenswrapper[4829]: I1002 07:35:25.731662 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerStarted","Data":"6adb581d4289c36239c944f3c373fd6919fb260b602e14f7f4a64e3c85102585"} Oct 02 07:35:26 crc kubenswrapper[4829]: I1002 07:35:26.103340 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 07:35:26 crc kubenswrapper[4829]: I1002 07:35:26.741958 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2bec1c3c-60b9-4776-93c0-3d51886b3552","Type":"ContainerStarted","Data":"848c6183f645770d3acabd4a413fa7174a85b6f0e2b0670869782186949ca7e3"} Oct 02 07:35:26 crc kubenswrapper[4829]: I1002 07:35:26.744757 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerStarted","Data":"d9cb8a3fed72a4f092b9151763405420aab78c8e51ffefedfefe41b0a5ef99e6"} Oct 02 07:35:26 crc kubenswrapper[4829]: I1002 07:35:26.745025 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:35:26 crc kubenswrapper[4829]: I1002 07:35:26.764217 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.790276802 podStartE2EDuration="6.764195859s" podCreationTimestamp="2025-10-02 07:35:20 +0000 UTC" firstStartedPulling="2025-10-02 07:35:22.390083589 +0000 UTC m=+1113.729731994" lastFinishedPulling="2025-10-02 07:35:26.364002656 +0000 UTC m=+1117.703651051" observedRunningTime="2025-10-02 07:35:26.761613946 +0000 UTC m=+1118.101262351" watchObservedRunningTime="2025-10-02 07:35:26.764195859 +0000 UTC m=+1118.103844264" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.350086 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.525212 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.596071 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6stp4\" (UniqueName: \"kubernetes.io/projected/a5724411-e0dd-4ab9-a898-0ecfbfe64385-kube-api-access-6stp4\") pod \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.596129 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-config\") pod \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.596187 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-httpd-config\") pod \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.596235 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-combined-ca-bundle\") pod \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.596351 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-ovndb-tls-certs\") pod \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\" (UID: \"a5724411-e0dd-4ab9-a898-0ecfbfe64385\") " Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.613415 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a5724411-e0dd-4ab9-a898-0ecfbfe64385" (UID: "a5724411-e0dd-4ab9-a898-0ecfbfe64385"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.613410 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5724411-e0dd-4ab9-a898-0ecfbfe64385-kube-api-access-6stp4" (OuterVolumeSpecName: "kube-api-access-6stp4") pod "a5724411-e0dd-4ab9-a898-0ecfbfe64385" (UID: "a5724411-e0dd-4ab9-a898-0ecfbfe64385"). InnerVolumeSpecName "kube-api-access-6stp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.666307 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5724411-e0dd-4ab9-a898-0ecfbfe64385" (UID: "a5724411-e0dd-4ab9-a898-0ecfbfe64385"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.676468 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-config" (OuterVolumeSpecName: "config") pod "a5724411-e0dd-4ab9-a898-0ecfbfe64385" (UID: "a5724411-e0dd-4ab9-a898-0ecfbfe64385"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.685986 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.696638 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a5724411-e0dd-4ab9-a898-0ecfbfe64385" (UID: "a5724411-e0dd-4ab9-a898-0ecfbfe64385"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.706452 4829 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.706489 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.706500 4829 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.706510 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6stp4\" (UniqueName: \"kubernetes.io/projected/a5724411-e0dd-4ab9-a898-0ecfbfe64385-kube-api-access-6stp4\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.706519 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5724411-e0dd-4ab9-a898-0ecfbfe64385-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.758465 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2bec1c3c-60b9-4776-93c0-3d51886b3552","Type":"ContainerStarted","Data":"6d8e252d82f1b441a3b0787afa2edc0fa59379f4d311680784ff47875aa76104"} Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.758515 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"2bec1c3c-60b9-4776-93c0-3d51886b3552","Type":"ContainerStarted","Data":"1fe251c2cad6c2b44a7b02c4881224df1031b33e716700b106a4a500f7539227"} Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.758671 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.769335 4829 generic.go:334] "Generic (PLEG): container finished" podID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerID="e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74" exitCode=0 Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.769850 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66cb4fcc98-c5psq" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.770302 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cb4fcc98-c5psq" event={"ID":"a5724411-e0dd-4ab9-a898-0ecfbfe64385","Type":"ContainerDied","Data":"e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74"} Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.770332 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66cb4fcc98-c5psq" event={"ID":"a5724411-e0dd-4ab9-a898-0ecfbfe64385","Type":"ContainerDied","Data":"465c00156d0bf7ba97844390214dcf7b02c0353e500693b5086803ddd3958989"} Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.770349 4829 scope.go:117] "RemoveContainer" containerID="d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.802311 4829 scope.go:117] "RemoveContainer" containerID="e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.807868 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.807849695 podStartE2EDuration="3.807849695s" podCreationTimestamp="2025-10-02 07:35:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:27.775187521 +0000 UTC m=+1119.114835926" watchObservedRunningTime="2025-10-02 07:35:27.807849695 +0000 UTC m=+1119.147498100" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.853410 4829 scope.go:117] "RemoveContainer" containerID="d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85" Oct 02 07:35:27 crc kubenswrapper[4829]: E1002 07:35:27.855718 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85\": container with ID starting with d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85 not found: ID does not exist" containerID="d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.855782 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85"} err="failed to get container status \"d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85\": rpc error: code = NotFound desc = could not find container \"d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85\": container with ID starting with d53df017344abb65bf187db34cf5c924ce8e420ae99e90af8bc1025794501e85 not found: ID does not exist" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.855808 4829 scope.go:117] "RemoveContainer" containerID="e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74" Oct 02 07:35:27 crc kubenswrapper[4829]: E1002 07:35:27.862666 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74\": container with ID starting with e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74 not found: ID does not exist" containerID="e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.862728 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74"} err="failed to get container status \"e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74\": rpc error: code = NotFound desc = could not find container \"e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74\": container with ID starting with e8c2dc8f991fd4e1d9df3c3f85897fcfd67126b662610ee3409af85aa7c32d74 not found: ID does not exist" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.894281 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66cb4fcc98-c5psq"] Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.908009 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66cb4fcc98-c5psq"] Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.945632 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5df99f8898-76z9s" Oct 02 07:35:27 crc kubenswrapper[4829]: I1002 07:35:27.972692 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.002467 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.016221 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7dcfb8bd44-ws7cq"] Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.016457 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7dcfb8bd44-ws7cq" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon-log" containerID="cri-o://9814c7fd97f5c331f361eb5c9bf44206b3cba730dc3c445f6f3835986ffb7bef" gracePeriod=30 Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.016596 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7dcfb8bd44-ws7cq" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" containerID="cri-o://dae3209bf1f7bb8956a52297b8d81605ad2f3b0ca6b90bc870590478cc3ac44f" gracePeriod=30 Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.045037 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.108584 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.347386 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.428450 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-9g6ss"] Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.428711 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerName="dnsmasq-dns" containerID="cri-o://0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15" gracePeriod=10 Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.793877 4829 generic.go:334] "Generic (PLEG): container finished" podID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerID="0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15" exitCode=0 Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.793942 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" event={"ID":"5052ce64-e1d7-429f-9256-f004ef04bc5a","Type":"ContainerDied","Data":"0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15"} Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.797772 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="cinder-scheduler" containerID="cri-o://80372dd323dd942281116dc27256c3c31e050ed8081300fbed8a5c7eae8110aa" gracePeriod=30 Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.798187 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="probe" containerID="cri-o://35b374e52770c3e6f5f3af1948ff51327535721d20cab9ef4bb9e57b03499f43" gracePeriod=30 Oct 02 07:35:28 crc kubenswrapper[4829]: E1002 07:35:28.815259 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5052ce64_e1d7_429f_9256_f004ef04bc5a.slice/crio-conmon-0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5052ce64_e1d7_429f_9256_f004ef04bc5a.slice/crio-0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:35:28 crc kubenswrapper[4829]: I1002 07:35:28.815727 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.133477 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.245260 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-swift-storage-0\") pod \"5052ce64-e1d7-429f-9256-f004ef04bc5a\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.245312 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwphc\" (UniqueName: \"kubernetes.io/projected/5052ce64-e1d7-429f-9256-f004ef04bc5a-kube-api-access-hwphc\") pod \"5052ce64-e1d7-429f-9256-f004ef04bc5a\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.245356 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-svc\") pod \"5052ce64-e1d7-429f-9256-f004ef04bc5a\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.245386 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-sb\") pod \"5052ce64-e1d7-429f-9256-f004ef04bc5a\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.245434 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-config\") pod \"5052ce64-e1d7-429f-9256-f004ef04bc5a\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.245496 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-nb\") pod \"5052ce64-e1d7-429f-9256-f004ef04bc5a\" (UID: \"5052ce64-e1d7-429f-9256-f004ef04bc5a\") " Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.254491 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5052ce64-e1d7-429f-9256-f004ef04bc5a-kube-api-access-hwphc" (OuterVolumeSpecName: "kube-api-access-hwphc") pod "5052ce64-e1d7-429f-9256-f004ef04bc5a" (UID: "5052ce64-e1d7-429f-9256-f004ef04bc5a"). InnerVolumeSpecName "kube-api-access-hwphc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.295632 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5052ce64-e1d7-429f-9256-f004ef04bc5a" (UID: "5052ce64-e1d7-429f-9256-f004ef04bc5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.310452 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5052ce64-e1d7-429f-9256-f004ef04bc5a" (UID: "5052ce64-e1d7-429f-9256-f004ef04bc5a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.319164 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-config" (OuterVolumeSpecName: "config") pod "5052ce64-e1d7-429f-9256-f004ef04bc5a" (UID: "5052ce64-e1d7-429f-9256-f004ef04bc5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.322892 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5052ce64-e1d7-429f-9256-f004ef04bc5a" (UID: "5052ce64-e1d7-429f-9256-f004ef04bc5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.326350 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5052ce64-e1d7-429f-9256-f004ef04bc5a" (UID: "5052ce64-e1d7-429f-9256-f004ef04bc5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.347533 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.347566 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwphc\" (UniqueName: \"kubernetes.io/projected/5052ce64-e1d7-429f-9256-f004ef04bc5a-kube-api-access-hwphc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.347575 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.347584 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.347595 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.347603 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5052ce64-e1d7-429f-9256-f004ef04bc5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.470862 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" path="/var/lib/kubelet/pods/a5724411-e0dd-4ab9-a898-0ecfbfe64385/volumes" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.808910 4829 generic.go:334] "Generic (PLEG): container finished" podID="e09c564a-17db-4723-a709-4fa0e897af52" containerID="35b374e52770c3e6f5f3af1948ff51327535721d20cab9ef4bb9e57b03499f43" exitCode=0 Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.808964 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e09c564a-17db-4723-a709-4fa0e897af52","Type":"ContainerDied","Data":"35b374e52770c3e6f5f3af1948ff51327535721d20cab9ef4bb9e57b03499f43"} Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.812373 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.812503 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-9g6ss" event={"ID":"5052ce64-e1d7-429f-9256-f004ef04bc5a","Type":"ContainerDied","Data":"9d366d589dfc4d9c501814df7cfe8b5d5f033da793dd12d0b72f056054caef21"} Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.812659 4829 scope.go:117] "RemoveContainer" containerID="0b7fae48f1bc67eea242c2447d8069b3529f527d3db25751b1a09ebed0571b15" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.855645 4829 scope.go:117] "RemoveContainer" containerID="97f6ef6f9592cc02d589748832968f9e172748f384d0646837adf5a09af6c799" Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.869095 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-9g6ss"] Oct 02 07:35:29 crc kubenswrapper[4829]: I1002 07:35:29.887346 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-9g6ss"] Oct 02 07:35:30 crc kubenswrapper[4829]: I1002 07:35:30.099442 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:30 crc kubenswrapper[4829]: I1002 07:35:30.389677 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.473435 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" path="/var/lib/kubelet/pods/5052ce64-e1d7-429f-9256-f004ef04bc5a/volumes" Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.506999 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7dcfb8bd44-ws7cq" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.771835 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.779925 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-995ff44cb-bcb2z" Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.871439 4829 generic.go:334] "Generic (PLEG): container finished" podID="e09c564a-17db-4723-a709-4fa0e897af52" containerID="80372dd323dd942281116dc27256c3c31e050ed8081300fbed8a5c7eae8110aa" exitCode=0 Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.871878 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e09c564a-17db-4723-a709-4fa0e897af52","Type":"ContainerDied","Data":"80372dd323dd942281116dc27256c3c31e050ed8081300fbed8a5c7eae8110aa"} Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.876808 4829 generic.go:334] "Generic (PLEG): container finished" podID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerID="dae3209bf1f7bb8956a52297b8d81605ad2f3b0ca6b90bc870590478cc3ac44f" exitCode=0 Oct 02 07:35:31 crc kubenswrapper[4829]: I1002 07:35:31.877775 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dcfb8bd44-ws7cq" event={"ID":"3a176a27-612f-499a-82c4-645d9cdd0b8b","Type":"ContainerDied","Data":"dae3209bf1f7bb8956a52297b8d81605ad2f3b0ca6b90bc870590478cc3ac44f"} Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.043812 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145381 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data\") pod \"e09c564a-17db-4723-a709-4fa0e897af52\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145430 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data-custom\") pod \"e09c564a-17db-4723-a709-4fa0e897af52\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145513 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-combined-ca-bundle\") pod \"e09c564a-17db-4723-a709-4fa0e897af52\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145558 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-scripts\") pod \"e09c564a-17db-4723-a709-4fa0e897af52\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145624 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e09c564a-17db-4723-a709-4fa0e897af52-etc-machine-id\") pod \"e09c564a-17db-4723-a709-4fa0e897af52\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145651 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzhtx\" (UniqueName: \"kubernetes.io/projected/e09c564a-17db-4723-a709-4fa0e897af52-kube-api-access-qzhtx\") pod \"e09c564a-17db-4723-a709-4fa0e897af52\" (UID: \"e09c564a-17db-4723-a709-4fa0e897af52\") " Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.145705 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e09c564a-17db-4723-a709-4fa0e897af52-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e09c564a-17db-4723-a709-4fa0e897af52" (UID: "e09c564a-17db-4723-a709-4fa0e897af52"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.146202 4829 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e09c564a-17db-4723-a709-4fa0e897af52-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.151011 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-scripts" (OuterVolumeSpecName: "scripts") pod "e09c564a-17db-4723-a709-4fa0e897af52" (UID: "e09c564a-17db-4723-a709-4fa0e897af52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.164362 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e09c564a-17db-4723-a709-4fa0e897af52" (UID: "e09c564a-17db-4723-a709-4fa0e897af52"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.171835 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e09c564a-17db-4723-a709-4fa0e897af52-kube-api-access-qzhtx" (OuterVolumeSpecName: "kube-api-access-qzhtx") pod "e09c564a-17db-4723-a709-4fa0e897af52" (UID: "e09c564a-17db-4723-a709-4fa0e897af52"). InnerVolumeSpecName "kube-api-access-qzhtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.219635 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e09c564a-17db-4723-a709-4fa0e897af52" (UID: "e09c564a-17db-4723-a709-4fa0e897af52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.247387 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzhtx\" (UniqueName: \"kubernetes.io/projected/e09c564a-17db-4723-a709-4fa0e897af52-kube-api-access-qzhtx\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.247419 4829 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.247427 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.247438 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.260057 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data" (OuterVolumeSpecName: "config-data") pod "e09c564a-17db-4723-a709-4fa0e897af52" (UID: "e09c564a-17db-4723-a709-4fa0e897af52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.349398 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09c564a-17db-4723-a709-4fa0e897af52-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.738493 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6b77fd5f6d-mmltx" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.889039 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.888970 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e09c564a-17db-4723-a709-4fa0e897af52","Type":"ContainerDied","Data":"638dab5bccff0224b3acbb7e49685978a3b4b6d4cab6c274cf1be43d35e8e5d9"} Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.889405 4829 scope.go:117] "RemoveContainer" containerID="35b374e52770c3e6f5f3af1948ff51327535721d20cab9ef4bb9e57b03499f43" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.922296 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.927393 4829 scope.go:117] "RemoveContainer" containerID="80372dd323dd942281116dc27256c3c31e050ed8081300fbed8a5c7eae8110aa" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.928725 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.980559 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:32 crc kubenswrapper[4829]: E1002 07:35:32.981433 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-api" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.981450 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-api" Oct 02 07:35:32 crc kubenswrapper[4829]: E1002 07:35:32.981471 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="cinder-scheduler" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.981479 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="cinder-scheduler" Oct 02 07:35:32 crc kubenswrapper[4829]: E1002 07:35:32.981509 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerName="dnsmasq-dns" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.981515 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerName="dnsmasq-dns" Oct 02 07:35:32 crc kubenswrapper[4829]: E1002 07:35:32.981551 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="probe" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.981556 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="probe" Oct 02 07:35:32 crc kubenswrapper[4829]: E1002 07:35:32.981573 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerName="init" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.981579 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerName="init" Oct 02 07:35:32 crc kubenswrapper[4829]: E1002 07:35:32.981602 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-httpd" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.981608 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-httpd" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.982061 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="probe" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.982089 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-api" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.982115 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5724411-e0dd-4ab9-a898-0ecfbfe64385" containerName="neutron-httpd" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.982133 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="5052ce64-e1d7-429f-9256-f004ef04bc5a" containerName="dnsmasq-dns" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.982155 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09c564a-17db-4723-a709-4fa0e897af52" containerName="cinder-scheduler" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.983983 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:35:32 crc kubenswrapper[4829]: I1002 07:35:32.987930 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.045632 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.062779 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vsmq\" (UniqueName: \"kubernetes.io/projected/5da522c4-7a44-46bc-b84e-eb230aad0bb7-kube-api-access-2vsmq\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.062882 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.062952 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.062981 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-config-data\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.063090 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-scripts\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.063112 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5da522c4-7a44-46bc-b84e-eb230aad0bb7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.164355 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-scripts\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.164410 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5da522c4-7a44-46bc-b84e-eb230aad0bb7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.164557 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vsmq\" (UniqueName: \"kubernetes.io/projected/5da522c4-7a44-46bc-b84e-eb230aad0bb7-kube-api-access-2vsmq\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.164599 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.164655 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.164684 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-config-data\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.166097 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5da522c4-7a44-46bc-b84e-eb230aad0bb7-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.169809 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-scripts\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.170903 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-config-data\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.173620 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.185455 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5da522c4-7a44-46bc-b84e-eb230aad0bb7-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.200390 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vsmq\" (UniqueName: \"kubernetes.io/projected/5da522c4-7a44-46bc-b84e-eb230aad0bb7-kube-api-access-2vsmq\") pod \"cinder-scheduler-0\" (UID: \"5da522c4-7a44-46bc-b84e-eb230aad0bb7\") " pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.300029 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.480187 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e09c564a-17db-4723-a709-4fa0e897af52" path="/var/lib/kubelet/pods/e09c564a-17db-4723-a709-4fa0e897af52/volumes" Oct 02 07:35:33 crc kubenswrapper[4829]: W1002 07:35:33.808150 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5da522c4_7a44_46bc_b84e_eb230aad0bb7.slice/crio-d61bd2a0f5403cd7cc286f94eb1cf84094505dffc383f90bd71d8244ff41a329 WatchSource:0}: Error finding container d61bd2a0f5403cd7cc286f94eb1cf84094505dffc383f90bd71d8244ff41a329: Status 404 returned error can't find the container with id d61bd2a0f5403cd7cc286f94eb1cf84094505dffc383f90bd71d8244ff41a329 Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.814588 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 07:35:33 crc kubenswrapper[4829]: I1002 07:35:33.935878 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5da522c4-7a44-46bc-b84e-eb230aad0bb7","Type":"ContainerStarted","Data":"d61bd2a0f5403cd7cc286f94eb1cf84094505dffc383f90bd71d8244ff41a329"} Oct 02 07:35:34 crc kubenswrapper[4829]: I1002 07:35:34.957265 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5da522c4-7a44-46bc-b84e-eb230aad0bb7","Type":"ContainerStarted","Data":"eb4593d39ba566abd717dc537d30ec89542c23c62087f9bb8ba8f07d50b7a99c"} Oct 02 07:35:35 crc kubenswrapper[4829]: I1002 07:35:35.968763 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5da522c4-7a44-46bc-b84e-eb230aad0bb7","Type":"ContainerStarted","Data":"4fe9ce37e927cf3dbedb01bc02d79ba6c152bb134ca1642d7403b77ae57854d1"} Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.004142 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.004125274 podStartE2EDuration="4.004125274s" podCreationTimestamp="2025-10-02 07:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:36.001659765 +0000 UTC m=+1127.341308180" watchObservedRunningTime="2025-10-02 07:35:36.004125274 +0000 UTC m=+1127.343773679" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.036692 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.037882 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.046883 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.047150 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.047602 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.047788 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-f897j" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.132587 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/859798c1-faae-46b4-83f2-0b9d18734236-openstack-config\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.132714 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dm6k\" (UniqueName: \"kubernetes.io/projected/859798c1-faae-46b4-83f2-0b9d18734236-kube-api-access-6dm6k\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.132809 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/859798c1-faae-46b4-83f2-0b9d18734236-openstack-config-secret\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.132838 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859798c1-faae-46b4-83f2-0b9d18734236-combined-ca-bundle\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.234507 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859798c1-faae-46b4-83f2-0b9d18734236-combined-ca-bundle\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.234619 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/859798c1-faae-46b4-83f2-0b9d18734236-openstack-config\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.234703 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dm6k\" (UniqueName: \"kubernetes.io/projected/859798c1-faae-46b4-83f2-0b9d18734236-kube-api-access-6dm6k\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.234752 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/859798c1-faae-46b4-83f2-0b9d18734236-openstack-config-secret\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.235675 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/859798c1-faae-46b4-83f2-0b9d18734236-openstack-config\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.242833 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/859798c1-faae-46b4-83f2-0b9d18734236-combined-ca-bundle\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.244366 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/859798c1-faae-46b4-83f2-0b9d18734236-openstack-config-secret\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.267663 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dm6k\" (UniqueName: \"kubernetes.io/projected/859798c1-faae-46b4-83f2-0b9d18734236-kube-api-access-6dm6k\") pod \"openstackclient\" (UID: \"859798c1-faae-46b4-83f2-0b9d18734236\") " pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.315199 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.356885 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.483833 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c7944ccd6-g5gh7" Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.619996 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b44f6ccc4-59gl5"] Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.620648 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b44f6ccc4-59gl5" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api-log" containerID="cri-o://f407cf8ffe266d1518e3a446c838a1c806736bf5fa0f3aba4b5c0165ec7741b2" gracePeriod=30 Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.620947 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7b44f6ccc4-59gl5" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api" containerID="cri-o://7d41ca2ed50d784ce82c011f41dac6aee40ce5935d199c629a92b077203e99f4" gracePeriod=30 Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.978828 4829 generic.go:334] "Generic (PLEG): container finished" podID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerID="f407cf8ffe266d1518e3a446c838a1c806736bf5fa0f3aba4b5c0165ec7741b2" exitCode=143 Oct 02 07:35:36 crc kubenswrapper[4829]: I1002 07:35:36.978910 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b44f6ccc4-59gl5" event={"ID":"f5eeddb3-db01-4916-94c8-a4914b521a0b","Type":"ContainerDied","Data":"f407cf8ffe266d1518e3a446c838a1c806736bf5fa0f3aba4b5c0165ec7741b2"} Oct 02 07:35:37 crc kubenswrapper[4829]: I1002 07:35:37.004189 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 07:35:37 crc kubenswrapper[4829]: I1002 07:35:37.807081 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 07:35:37 crc kubenswrapper[4829]: I1002 07:35:37.999371 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"859798c1-faae-46b4-83f2-0b9d18734236","Type":"ContainerStarted","Data":"ce7942601a08babc39cbbbb3a3e94268a6ee32b50858426486f7c0a6c2c45ea0"} Oct 02 07:35:38 crc kubenswrapper[4829]: I1002 07:35:38.300582 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.728916 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-84b89d7c8c-f2lgb"] Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.732181 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.736246 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.736359 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.736487 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.791361 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-84b89d7c8c-f2lgb"] Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802706 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-792jz\" (UniqueName: \"kubernetes.io/projected/395b490d-fbc2-403d-ac35-9541329363c2-kube-api-access-792jz\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802752 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-combined-ca-bundle\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802834 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395b490d-fbc2-403d-ac35-9541329363c2-run-httpd\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802851 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395b490d-fbc2-403d-ac35-9541329363c2-log-httpd\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802867 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-config-data\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802909 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-internal-tls-certs\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802960 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/395b490d-fbc2-403d-ac35-9541329363c2-etc-swift\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.802979 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-public-tls-certs\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.817379 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b44f6ccc4-59gl5" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": read tcp 10.217.0.2:41460->10.217.0.179:9311: read: connection reset by peer" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.817433 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b44f6ccc4-59gl5" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": read tcp 10.217.0.2:41462->10.217.0.179:9311: read: connection reset by peer" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909297 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395b490d-fbc2-403d-ac35-9541329363c2-run-httpd\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909343 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395b490d-fbc2-403d-ac35-9541329363c2-log-httpd\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909374 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-config-data\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909426 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-internal-tls-certs\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909480 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/395b490d-fbc2-403d-ac35-9541329363c2-etc-swift\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909500 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-public-tls-certs\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909541 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-combined-ca-bundle\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.909560 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-792jz\" (UniqueName: \"kubernetes.io/projected/395b490d-fbc2-403d-ac35-9541329363c2-kube-api-access-792jz\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.910323 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395b490d-fbc2-403d-ac35-9541329363c2-run-httpd\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.910545 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/395b490d-fbc2-403d-ac35-9541329363c2-log-httpd\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.924640 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-combined-ca-bundle\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.925318 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/395b490d-fbc2-403d-ac35-9541329363c2-etc-swift\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.925455 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-public-tls-certs\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.925572 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-config-data\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.930946 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/395b490d-fbc2-403d-ac35-9541329363c2-internal-tls-certs\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:39 crc kubenswrapper[4829]: I1002 07:35:39.943740 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-792jz\" (UniqueName: \"kubernetes.io/projected/395b490d-fbc2-403d-ac35-9541329363c2-kube-api-access-792jz\") pod \"swift-proxy-84b89d7c8c-f2lgb\" (UID: \"395b490d-fbc2-403d-ac35-9541329363c2\") " pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.049404 4829 generic.go:334] "Generic (PLEG): container finished" podID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerID="7d41ca2ed50d784ce82c011f41dac6aee40ce5935d199c629a92b077203e99f4" exitCode=0 Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.049629 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b44f6ccc4-59gl5" event={"ID":"f5eeddb3-db01-4916-94c8-a4914b521a0b","Type":"ContainerDied","Data":"7d41ca2ed50d784ce82c011f41dac6aee40ce5935d199c629a92b077203e99f4"} Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.080875 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.425340 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.493947 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.499328 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="proxy-httpd" containerID="cri-o://d9cb8a3fed72a4f092b9151763405420aab78c8e51ffefedfefe41b0a5ef99e6" gracePeriod=30 Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.499944 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-notification-agent" containerID="cri-o://fa5d98bb57809fcf6500388558873874ef292c270da6cfcdffc7088cba73cfc2" gracePeriod=30 Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.500010 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="sg-core" containerID="cri-o://6adb581d4289c36239c944f3c373fd6919fb260b602e14f7f4a64e3c85102585" gracePeriod=30 Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.500055 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-central-agent" containerID="cri-o://f82dba84ad8187641e0688eef29bb60204eb0d65f94df93c4f36e4d52426c8a6" gracePeriod=30 Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.525068 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqd6z\" (UniqueName: \"kubernetes.io/projected/f5eeddb3-db01-4916-94c8-a4914b521a0b-kube-api-access-jqd6z\") pod \"f5eeddb3-db01-4916-94c8-a4914b521a0b\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.525117 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.529633 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-combined-ca-bundle\") pod \"f5eeddb3-db01-4916-94c8-a4914b521a0b\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.529744 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data-custom\") pod \"f5eeddb3-db01-4916-94c8-a4914b521a0b\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.529781 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data\") pod \"f5eeddb3-db01-4916-94c8-a4914b521a0b\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.529842 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5eeddb3-db01-4916-94c8-a4914b521a0b-logs\") pod \"f5eeddb3-db01-4916-94c8-a4914b521a0b\" (UID: \"f5eeddb3-db01-4916-94c8-a4914b521a0b\") " Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.533735 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5eeddb3-db01-4916-94c8-a4914b521a0b-logs" (OuterVolumeSpecName: "logs") pod "f5eeddb3-db01-4916-94c8-a4914b521a0b" (UID: "f5eeddb3-db01-4916-94c8-a4914b521a0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.534467 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5eeddb3-db01-4916-94c8-a4914b521a0b-kube-api-access-jqd6z" (OuterVolumeSpecName: "kube-api-access-jqd6z") pod "f5eeddb3-db01-4916-94c8-a4914b521a0b" (UID: "f5eeddb3-db01-4916-94c8-a4914b521a0b"). InnerVolumeSpecName "kube-api-access-jqd6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.535219 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f5eeddb3-db01-4916-94c8-a4914b521a0b" (UID: "f5eeddb3-db01-4916-94c8-a4914b521a0b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.577812 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5eeddb3-db01-4916-94c8-a4914b521a0b" (UID: "f5eeddb3-db01-4916-94c8-a4914b521a0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.620754 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data" (OuterVolumeSpecName: "config-data") pod "f5eeddb3-db01-4916-94c8-a4914b521a0b" (UID: "f5eeddb3-db01-4916-94c8-a4914b521a0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.633106 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.633144 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5eeddb3-db01-4916-94c8-a4914b521a0b-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.633156 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqd6z\" (UniqueName: \"kubernetes.io/projected/f5eeddb3-db01-4916-94c8-a4914b521a0b-kube-api-access-jqd6z\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.633174 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.633184 4829 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5eeddb3-db01-4916-94c8-a4914b521a0b-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:40 crc kubenswrapper[4829]: I1002 07:35:40.736892 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-84b89d7c8c-f2lgb"] Oct 02 07:35:40 crc kubenswrapper[4829]: W1002 07:35:40.748175 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod395b490d_fbc2_403d_ac35_9541329363c2.slice/crio-a53b1b50b652e62ad4b2fa89633143eb8759bc5336e99b69975c3b65b7b9e8b9 WatchSource:0}: Error finding container a53b1b50b652e62ad4b2fa89633143eb8759bc5336e99b69975c3b65b7b9e8b9: Status 404 returned error can't find the container with id a53b1b50b652e62ad4b2fa89633143eb8759bc5336e99b69975c3b65b7b9e8b9 Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.061638 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" event={"ID":"395b490d-fbc2-403d-ac35-9541329363c2","Type":"ContainerStarted","Data":"a53b1b50b652e62ad4b2fa89633143eb8759bc5336e99b69975c3b65b7b9e8b9"} Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.063683 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b44f6ccc4-59gl5" event={"ID":"f5eeddb3-db01-4916-94c8-a4914b521a0b","Type":"ContainerDied","Data":"efc543cb030f8325744e88a64a8e4b333ad1f1408cc53b2a8ebf5a5c4b899919"} Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.063734 4829 scope.go:117] "RemoveContainer" containerID="7d41ca2ed50d784ce82c011f41dac6aee40ce5935d199c629a92b077203e99f4" Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.063734 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b44f6ccc4-59gl5" Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.074045 4829 generic.go:334] "Generic (PLEG): container finished" podID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerID="d9cb8a3fed72a4f092b9151763405420aab78c8e51ffefedfefe41b0a5ef99e6" exitCode=0 Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.074074 4829 generic.go:334] "Generic (PLEG): container finished" podID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerID="6adb581d4289c36239c944f3c373fd6919fb260b602e14f7f4a64e3c85102585" exitCode=2 Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.074081 4829 generic.go:334] "Generic (PLEG): container finished" podID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerID="f82dba84ad8187641e0688eef29bb60204eb0d65f94df93c4f36e4d52426c8a6" exitCode=0 Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.074100 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerDied","Data":"d9cb8a3fed72a4f092b9151763405420aab78c8e51ffefedfefe41b0a5ef99e6"} Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.074125 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerDied","Data":"6adb581d4289c36239c944f3c373fd6919fb260b602e14f7f4a64e3c85102585"} Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.074134 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerDied","Data":"f82dba84ad8187641e0688eef29bb60204eb0d65f94df93c4f36e4d52426c8a6"} Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.096113 4829 scope.go:117] "RemoveContainer" containerID="f407cf8ffe266d1518e3a446c838a1c806736bf5fa0f3aba4b5c0165ec7741b2" Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.101028 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7b44f6ccc4-59gl5"] Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.108510 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7b44f6ccc4-59gl5"] Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.472512 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" path="/var/lib/kubelet/pods/f5eeddb3-db01-4916-94c8-a4914b521a0b/volumes" Oct 02 07:35:41 crc kubenswrapper[4829]: I1002 07:35:41.506787 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7dcfb8bd44-ws7cq" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 02 07:35:42 crc kubenswrapper[4829]: I1002 07:35:42.092033 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" event={"ID":"395b490d-fbc2-403d-ac35-9541329363c2","Type":"ContainerStarted","Data":"c773cb458ff3fb25a85cf6b4224f072e8d83432d894a600666b0b97fad4bd93d"} Oct 02 07:35:42 crc kubenswrapper[4829]: I1002 07:35:42.092075 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" event={"ID":"395b490d-fbc2-403d-ac35-9541329363c2","Type":"ContainerStarted","Data":"f43985cdc9c38cb29ffef84b031a744c76ca100ce360ded339072508249ec097"} Oct 02 07:35:42 crc kubenswrapper[4829]: I1002 07:35:42.093461 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:42 crc kubenswrapper[4829]: I1002 07:35:42.093485 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:42 crc kubenswrapper[4829]: I1002 07:35:42.129653 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" podStartSLOduration=3.129629377 podStartE2EDuration="3.129629377s" podCreationTimestamp="2025-10-02 07:35:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:42.113800126 +0000 UTC m=+1133.453448531" watchObservedRunningTime="2025-10-02 07:35:42.129629377 +0000 UTC m=+1133.469277782" Oct 02 07:35:43 crc kubenswrapper[4829]: I1002 07:35:43.120516 4829 generic.go:334] "Generic (PLEG): container finished" podID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerID="fa5d98bb57809fcf6500388558873874ef292c270da6cfcdffc7088cba73cfc2" exitCode=0 Oct 02 07:35:43 crc kubenswrapper[4829]: I1002 07:35:43.121617 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerDied","Data":"fa5d98bb57809fcf6500388558873874ef292c270da6cfcdffc7088cba73cfc2"} Oct 02 07:35:43 crc kubenswrapper[4829]: I1002 07:35:43.531849 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.414272 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-d754t"] Oct 02 07:35:44 crc kubenswrapper[4829]: E1002 07:35:44.414640 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api-log" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.414652 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api-log" Oct 02 07:35:44 crc kubenswrapper[4829]: E1002 07:35:44.414683 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.414689 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.414861 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api-log" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.414876 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5eeddb3-db01-4916-94c8-a4914b521a0b" containerName="barbican-api" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.423394 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.428789 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-d754t"] Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.520706 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-sqbkw"] Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.522296 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.527883 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sqbkw"] Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.612430 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pnfn\" (UniqueName: \"kubernetes.io/projected/59c48271-56c9-483d-8db7-0e9cb785e2d2-kube-api-access-9pnfn\") pod \"nova-api-db-create-d754t\" (UID: \"59c48271-56c9-483d-8db7-0e9cb785e2d2\") " pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.627430 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-w9kh7"] Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.629052 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.635250 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-w9kh7"] Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.715345 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pnfn\" (UniqueName: \"kubernetes.io/projected/59c48271-56c9-483d-8db7-0e9cb785e2d2-kube-api-access-9pnfn\") pod \"nova-api-db-create-d754t\" (UID: \"59c48271-56c9-483d-8db7-0e9cb785e2d2\") " pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.715491 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8kw5\" (UniqueName: \"kubernetes.io/projected/c5be0189-8767-4668-b280-7cefa2f9237d-kube-api-access-b8kw5\") pod \"nova-cell0-db-create-sqbkw\" (UID: \"c5be0189-8767-4668-b280-7cefa2f9237d\") " pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.740459 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pnfn\" (UniqueName: \"kubernetes.io/projected/59c48271-56c9-483d-8db7-0e9cb785e2d2-kube-api-access-9pnfn\") pod \"nova-api-db-create-d754t\" (UID: \"59c48271-56c9-483d-8db7-0e9cb785e2d2\") " pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.751995 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.818416 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnmrk\" (UniqueName: \"kubernetes.io/projected/4509f44c-18ac-449a-9b64-619c7c304e16-kube-api-access-jnmrk\") pod \"nova-cell1-db-create-w9kh7\" (UID: \"4509f44c-18ac-449a-9b64-619c7c304e16\") " pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.818546 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8kw5\" (UniqueName: \"kubernetes.io/projected/c5be0189-8767-4668-b280-7cefa2f9237d-kube-api-access-b8kw5\") pod \"nova-cell0-db-create-sqbkw\" (UID: \"c5be0189-8767-4668-b280-7cefa2f9237d\") " pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.842043 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8kw5\" (UniqueName: \"kubernetes.io/projected/c5be0189-8767-4668-b280-7cefa2f9237d-kube-api-access-b8kw5\") pod \"nova-cell0-db-create-sqbkw\" (UID: \"c5be0189-8767-4668-b280-7cefa2f9237d\") " pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.842565 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.920482 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnmrk\" (UniqueName: \"kubernetes.io/projected/4509f44c-18ac-449a-9b64-619c7c304e16-kube-api-access-jnmrk\") pod \"nova-cell1-db-create-w9kh7\" (UID: \"4509f44c-18ac-449a-9b64-619c7c304e16\") " pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.947007 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnmrk\" (UniqueName: \"kubernetes.io/projected/4509f44c-18ac-449a-9b64-619c7c304e16-kube-api-access-jnmrk\") pod \"nova-cell1-db-create-w9kh7\" (UID: \"4509f44c-18ac-449a-9b64-619c7c304e16\") " pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:44 crc kubenswrapper[4829]: I1002 07:35:44.956164 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:47 crc kubenswrapper[4829]: I1002 07:35:47.200983 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:35:47 crc kubenswrapper[4829]: I1002 07:35:47.201609 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-log" containerID="cri-o://5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f" gracePeriod=30 Oct 02 07:35:47 crc kubenswrapper[4829]: I1002 07:35:47.201672 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-httpd" containerID="cri-o://54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc" gracePeriod=30 Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.225065 4829 generic.go:334] "Generic (PLEG): container finished" podID="8104fa43-766f-4747-bfde-8efb050e4629" containerID="5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f" exitCode=143 Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.225416 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8104fa43-766f-4747-bfde-8efb050e4629","Type":"ContainerDied","Data":"5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f"} Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.508180 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.699108 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-w9kh7"] Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701100 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8hw\" (UniqueName: \"kubernetes.io/projected/f45b611a-62d2-4c65-8a5f-8557b05da53a-kube-api-access-ws8hw\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701138 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-combined-ca-bundle\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701194 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-run-httpd\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701219 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-config-data\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701300 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-log-httpd\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701373 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-sg-core-conf-yaml\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.701396 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-scripts\") pod \"f45b611a-62d2-4c65-8a5f-8557b05da53a\" (UID: \"f45b611a-62d2-4c65-8a5f-8557b05da53a\") " Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.702241 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.702364 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.714173 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-scripts" (OuterVolumeSpecName: "scripts") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.714403 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45b611a-62d2-4c65-8a5f-8557b05da53a-kube-api-access-ws8hw" (OuterVolumeSpecName: "kube-api-access-ws8hw") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "kube-api-access-ws8hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.741298 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.814394 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.823783 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.823804 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.823818 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ws8hw\" (UniqueName: \"kubernetes.io/projected/f45b611a-62d2-4c65-8a5f-8557b05da53a-kube-api-access-ws8hw\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.823830 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f45b611a-62d2-4c65-8a5f-8557b05da53a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.834341 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-config-data" (OuterVolumeSpecName: "config-data") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.837333 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-sqbkw"] Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.848728 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-d754t"] Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.859047 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f45b611a-62d2-4c65-8a5f-8557b05da53a" (UID: "f45b611a-62d2-4c65-8a5f-8557b05da53a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.925426 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:48 crc kubenswrapper[4829]: I1002 07:35:48.925792 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45b611a-62d2-4c65-8a5f-8557b05da53a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.235165 4829 generic.go:334] "Generic (PLEG): container finished" podID="59c48271-56c9-483d-8db7-0e9cb785e2d2" containerID="1d3af3a6768c0582737fd2ee09f23f7028a593a6df44b6f0afe97a463949f90c" exitCode=0 Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.235240 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d754t" event={"ID":"59c48271-56c9-483d-8db7-0e9cb785e2d2","Type":"ContainerDied","Data":"1d3af3a6768c0582737fd2ee09f23f7028a593a6df44b6f0afe97a463949f90c"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.235266 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d754t" event={"ID":"59c48271-56c9-483d-8db7-0e9cb785e2d2","Type":"ContainerStarted","Data":"2b5fdcc8685de361cc7cba173fa682ec7c59a40141e6267a291662f3b304f071"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.239889 4829 generic.go:334] "Generic (PLEG): container finished" podID="c5be0189-8767-4668-b280-7cefa2f9237d" containerID="144cc00dadc81fb60e62fb7d52c52e4cacad8f37f9d76931e3e67f9f408d84a3" exitCode=0 Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.239953 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sqbkw" event={"ID":"c5be0189-8767-4668-b280-7cefa2f9237d","Type":"ContainerDied","Data":"144cc00dadc81fb60e62fb7d52c52e4cacad8f37f9d76931e3e67f9f408d84a3"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.239980 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sqbkw" event={"ID":"c5be0189-8767-4668-b280-7cefa2f9237d","Type":"ContainerStarted","Data":"baa757e7982ba72a253ed5cff319505311c876758eac1e3b28574894d272fb85"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.246188 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f45b611a-62d2-4c65-8a5f-8557b05da53a","Type":"ContainerDied","Data":"09869c638bfc07a67dc444c439b9ff6464cdc1ed8f57c08a97e6af6f6ae32439"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.246241 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.246280 4829 scope.go:117] "RemoveContainer" containerID="d9cb8a3fed72a4f092b9151763405420aab78c8e51ffefedfefe41b0a5ef99e6" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.249898 4829 generic.go:334] "Generic (PLEG): container finished" podID="4509f44c-18ac-449a-9b64-619c7c304e16" containerID="1d7e5fded6980d5fdaa53714f64a89ae74e49847ed59fcecd6a4f4a60686d2fa" exitCode=0 Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.249974 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w9kh7" event={"ID":"4509f44c-18ac-449a-9b64-619c7c304e16","Type":"ContainerDied","Data":"1d7e5fded6980d5fdaa53714f64a89ae74e49847ed59fcecd6a4f4a60686d2fa"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.250003 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w9kh7" event={"ID":"4509f44c-18ac-449a-9b64-619c7c304e16","Type":"ContainerStarted","Data":"cdca9b6c86424a01afa78bc30588a1624c68f9b3dc2df803721a48bde08457e2"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.252093 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"859798c1-faae-46b4-83f2-0b9d18734236","Type":"ContainerStarted","Data":"95c2df9c3b97af62ecffe69b31364d2596d1f9ec745ece12782159aaa0c636ef"} Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.271831 4829 scope.go:117] "RemoveContainer" containerID="6adb581d4289c36239c944f3c373fd6919fb260b602e14f7f4a64e3c85102585" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.277355 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.138582418 podStartE2EDuration="13.277318912s" podCreationTimestamp="2025-10-02 07:35:36 +0000 UTC" firstStartedPulling="2025-10-02 07:35:37.012679098 +0000 UTC m=+1128.352327503" lastFinishedPulling="2025-10-02 07:35:48.151415592 +0000 UTC m=+1139.491063997" observedRunningTime="2025-10-02 07:35:49.267825025 +0000 UTC m=+1140.607473440" watchObservedRunningTime="2025-10-02 07:35:49.277318912 +0000 UTC m=+1140.616967317" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.302205 4829 scope.go:117] "RemoveContainer" containerID="fa5d98bb57809fcf6500388558873874ef292c270da6cfcdffc7088cba73cfc2" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.349593 4829 scope.go:117] "RemoveContainer" containerID="f82dba84ad8187641e0688eef29bb60204eb0d65f94df93c4f36e4d52426c8a6" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.372375 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.401083 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416022 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:49 crc kubenswrapper[4829]: E1002 07:35:49.416547 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-notification-agent" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416570 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-notification-agent" Oct 02 07:35:49 crc kubenswrapper[4829]: E1002 07:35:49.416598 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="proxy-httpd" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416606 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="proxy-httpd" Oct 02 07:35:49 crc kubenswrapper[4829]: E1002 07:35:49.416641 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="sg-core" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416649 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="sg-core" Oct 02 07:35:49 crc kubenswrapper[4829]: E1002 07:35:49.416672 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-central-agent" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416680 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-central-agent" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416898 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-notification-agent" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416917 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="sg-core" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416930 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="proxy-httpd" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.416954 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" containerName="ceilometer-central-agent" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.424911 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.427266 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.427855 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.436842 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.449327 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.449863 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-log" containerID="cri-o://41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b" gracePeriod=30 Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.450174 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-httpd" containerID="cri-o://6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62" gracePeriod=30 Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.493328 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f45b611a-62d2-4c65-8a5f-8557b05da53a" path="/var/lib/kubelet/pods/f45b611a-62d2-4c65-8a5f-8557b05da53a/volumes" Oct 02 07:35:49 crc kubenswrapper[4829]: E1002 07:35:49.513314 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59c48271_56c9_483d_8db7_0e9cb785e2d2.slice/crio-conmon-1d3af3a6768c0582737fd2ee09f23f7028a593a6df44b6f0afe97a463949f90c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf45b611a_62d2_4c65_8a5f_8557b05da53a.slice/crio-09869c638bfc07a67dc444c439b9ff6464cdc1ed8f57c08a97e6af6f6ae32439\": RecentStats: unable to find data in memory cache]" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537083 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-scripts\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537148 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537332 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-log-httpd\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537376 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537403 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfpwx\" (UniqueName: \"kubernetes.io/projected/4ba3b403-2a6c-4db1-add6-010fbd41728b-kube-api-access-qfpwx\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537425 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-run-httpd\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.537759 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-config-data\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.639807 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-scripts\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.639894 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.639963 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-log-httpd\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.639990 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.640011 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-run-httpd\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.640033 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfpwx\" (UniqueName: \"kubernetes.io/projected/4ba3b403-2a6c-4db1-add6-010fbd41728b-kube-api-access-qfpwx\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.640099 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-config-data\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.641115 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-log-httpd\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.642717 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-run-httpd\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.646839 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-config-data\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.647381 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-scripts\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.649641 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.656062 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.658423 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfpwx\" (UniqueName: \"kubernetes.io/projected/4ba3b403-2a6c-4db1-add6-010fbd41728b-kube-api-access-qfpwx\") pod \"ceilometer-0\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " pod="openstack/ceilometer-0" Oct 02 07:35:49 crc kubenswrapper[4829]: I1002 07:35:49.751440 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.087030 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.090599 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-84b89d7c8c-f2lgb" Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.262402 4829 generic.go:334] "Generic (PLEG): container finished" podID="63630a30-0c56-4bdb-b259-984b81733bcc" containerID="41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b" exitCode=143 Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.262722 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63630a30-0c56-4bdb-b259-984b81733bcc","Type":"ContainerDied","Data":"41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b"} Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.341656 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.777519 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.927813 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.968449 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pnfn\" (UniqueName: \"kubernetes.io/projected/59c48271-56c9-483d-8db7-0e9cb785e2d2-kube-api-access-9pnfn\") pod \"59c48271-56c9-483d-8db7-0e9cb785e2d2\" (UID: \"59c48271-56c9-483d-8db7-0e9cb785e2d2\") " Oct 02 07:35:50 crc kubenswrapper[4829]: I1002 07:35:50.982639 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59c48271-56c9-483d-8db7-0e9cb785e2d2-kube-api-access-9pnfn" (OuterVolumeSpecName: "kube-api-access-9pnfn") pod "59c48271-56c9-483d-8db7-0e9cb785e2d2" (UID: "59c48271-56c9-483d-8db7-0e9cb785e2d2"). InnerVolumeSpecName "kube-api-access-9pnfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.064517 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.071658 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnmrk\" (UniqueName: \"kubernetes.io/projected/4509f44c-18ac-449a-9b64-619c7c304e16-kube-api-access-jnmrk\") pod \"4509f44c-18ac-449a-9b64-619c7c304e16\" (UID: \"4509f44c-18ac-449a-9b64-619c7c304e16\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.072899 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pnfn\" (UniqueName: \"kubernetes.io/projected/59c48271-56c9-483d-8db7-0e9cb785e2d2-kube-api-access-9pnfn\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.074912 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4509f44c-18ac-449a-9b64-619c7c304e16-kube-api-access-jnmrk" (OuterVolumeSpecName: "kube-api-access-jnmrk") pod "4509f44c-18ac-449a-9b64-619c7c304e16" (UID: "4509f44c-18ac-449a-9b64-619c7c304e16"). InnerVolumeSpecName "kube-api-access-jnmrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.174614 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8kw5\" (UniqueName: \"kubernetes.io/projected/c5be0189-8767-4668-b280-7cefa2f9237d-kube-api-access-b8kw5\") pod \"c5be0189-8767-4668-b280-7cefa2f9237d\" (UID: \"c5be0189-8767-4668-b280-7cefa2f9237d\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.175605 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnmrk\" (UniqueName: \"kubernetes.io/projected/4509f44c-18ac-449a-9b64-619c7c304e16-kube-api-access-jnmrk\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.182358 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5be0189-8767-4668-b280-7cefa2f9237d-kube-api-access-b8kw5" (OuterVolumeSpecName: "kube-api-access-b8kw5") pod "c5be0189-8767-4668-b280-7cefa2f9237d" (UID: "c5be0189-8767-4668-b280-7cefa2f9237d"). InnerVolumeSpecName "kube-api-access-b8kw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.239705 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.275400 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerStarted","Data":"429ab66419b3f97cf2f9fdca2437d8575f4b1f168516a74278dd206482f45a48"} Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.276992 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8kw5\" (UniqueName: \"kubernetes.io/projected/c5be0189-8767-4668-b280-7cefa2f9237d-kube-api-access-b8kw5\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.278036 4829 generic.go:334] "Generic (PLEG): container finished" podID="8104fa43-766f-4747-bfde-8efb050e4629" containerID="54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc" exitCode=0 Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.278069 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8104fa43-766f-4747-bfde-8efb050e4629","Type":"ContainerDied","Data":"54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc"} Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.278093 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.278107 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8104fa43-766f-4747-bfde-8efb050e4629","Type":"ContainerDied","Data":"ade4731835022ced7e89b38c9645bcdbba6daa3cc3bd28b5fc1e545a473f0874"} Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.278124 4829 scope.go:117] "RemoveContainer" containerID="54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.285573 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-d754t" event={"ID":"59c48271-56c9-483d-8db7-0e9cb785e2d2","Type":"ContainerDied","Data":"2b5fdcc8685de361cc7cba173fa682ec7c59a40141e6267a291662f3b304f071"} Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.285632 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b5fdcc8685de361cc7cba173fa682ec7c59a40141e6267a291662f3b304f071" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.285734 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-d754t" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.287763 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-sqbkw" event={"ID":"c5be0189-8767-4668-b280-7cefa2f9237d","Type":"ContainerDied","Data":"baa757e7982ba72a253ed5cff319505311c876758eac1e3b28574894d272fb85"} Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.287799 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="baa757e7982ba72a253ed5cff319505311c876758eac1e3b28574894d272fb85" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.287798 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-sqbkw" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.289064 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w9kh7" event={"ID":"4509f44c-18ac-449a-9b64-619c7c304e16","Type":"ContainerDied","Data":"cdca9b6c86424a01afa78bc30588a1624c68f9b3dc2df803721a48bde08457e2"} Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.289085 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdca9b6c86424a01afa78bc30588a1624c68f9b3dc2df803721a48bde08457e2" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.289149 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w9kh7" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.304282 4829 scope.go:117] "RemoveContainer" containerID="5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.337822 4829 scope.go:117] "RemoveContainer" containerID="54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc" Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.338332 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc\": container with ID starting with 54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc not found: ID does not exist" containerID="54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.338368 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc"} err="failed to get container status \"54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc\": rpc error: code = NotFound desc = could not find container \"54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc\": container with ID starting with 54e5a147eb22dda0921c29e746379f7c1d48ac5c132629854acef42243ed58cc not found: ID does not exist" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.338389 4829 scope.go:117] "RemoveContainer" containerID="5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f" Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.339060 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f\": container with ID starting with 5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f not found: ID does not exist" containerID="5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.339091 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f"} err="failed to get container status \"5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f\": rpc error: code = NotFound desc = could not find container \"5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f\": container with ID starting with 5f079920991260193717c756d8d768cfcdee5446a6586777e190b663c6481d2f not found: ID does not exist" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378567 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-config-data\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378649 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378740 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw8lj\" (UniqueName: \"kubernetes.io/projected/8104fa43-766f-4747-bfde-8efb050e4629-kube-api-access-nw8lj\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378856 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-httpd-run\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378893 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-combined-ca-bundle\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378935 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-scripts\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.378975 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-logs\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.379012 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-public-tls-certs\") pod \"8104fa43-766f-4747-bfde-8efb050e4629\" (UID: \"8104fa43-766f-4747-bfde-8efb050e4629\") " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.379424 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.379971 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-logs" (OuterVolumeSpecName: "logs") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.384348 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-scripts" (OuterVolumeSpecName: "scripts") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.384379 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.385450 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8104fa43-766f-4747-bfde-8efb050e4629-kube-api-access-nw8lj" (OuterVolumeSpecName: "kube-api-access-nw8lj") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "kube-api-access-nw8lj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.415504 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.451948 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-config-data" (OuterVolumeSpecName: "config-data") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.452314 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8104fa43-766f-4747-bfde-8efb050e4629" (UID: "8104fa43-766f-4747-bfde-8efb050e4629"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.478191 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480624 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480651 4829 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480662 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480686 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480704 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw8lj\" (UniqueName: \"kubernetes.io/projected/8104fa43-766f-4747-bfde-8efb050e4629-kube-api-access-nw8lj\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480715 4829 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8104fa43-766f-4747-bfde-8efb050e4629-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480729 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.480737 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8104fa43-766f-4747-bfde-8efb050e4629-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.508727 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7dcfb8bd44-ws7cq" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.508832 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.509763 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.582422 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.600655 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.609362 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.625297 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.625762 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-log" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.625786 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-log" Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.625805 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4509f44c-18ac-449a-9b64-619c7c304e16" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.625813 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4509f44c-18ac-449a-9b64-619c7c304e16" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.625827 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c48271-56c9-483d-8db7-0e9cb785e2d2" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.625833 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c48271-56c9-483d-8db7-0e9cb785e2d2" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.625846 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5be0189-8767-4668-b280-7cefa2f9237d" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.625852 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5be0189-8767-4668-b280-7cefa2f9237d" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: E1002 07:35:51.625886 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-httpd" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.625892 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-httpd" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.626060 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-httpd" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.626084 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8104fa43-766f-4747-bfde-8efb050e4629" containerName="glance-log" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.626095 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4509f44c-18ac-449a-9b64-619c7c304e16" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.626109 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5be0189-8767-4668-b280-7cefa2f9237d" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.626121 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="59c48271-56c9-483d-8db7-0e9cb785e2d2" containerName="mariadb-database-create" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.627193 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.629532 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.630667 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.652332 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785007 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4lqn\" (UniqueName: \"kubernetes.io/projected/3f017e85-e428-48e7-9b00-705ca534019d-kube-api-access-z4lqn\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785198 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785336 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785566 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785617 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f017e85-e428-48e7-9b00-705ca534019d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785747 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785789 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f017e85-e428-48e7-9b00-705ca534019d-logs\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.785937 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887619 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887677 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887771 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f017e85-e428-48e7-9b00-705ca534019d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887825 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887853 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f017e85-e428-48e7-9b00-705ca534019d-logs\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.888389 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.888494 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4lqn\" (UniqueName: \"kubernetes.io/projected/3f017e85-e428-48e7-9b00-705ca534019d-kube-api-access-z4lqn\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.888314 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f017e85-e428-48e7-9b00-705ca534019d-logs\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.887996 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.888314 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3f017e85-e428-48e7-9b00-705ca534019d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.893868 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.894040 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.894724 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.897840 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f017e85-e428-48e7-9b00-705ca534019d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.906493 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4lqn\" (UniqueName: \"kubernetes.io/projected/3f017e85-e428-48e7-9b00-705ca534019d-kube-api-access-z4lqn\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.930732 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"3f017e85-e428-48e7-9b00-705ca534019d\") " pod="openstack/glance-default-external-api-0" Oct 02 07:35:51 crc kubenswrapper[4829]: I1002 07:35:51.945036 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 07:35:52 crc kubenswrapper[4829]: I1002 07:35:52.301852 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerStarted","Data":"2f99639131442bb6a98dc44bcbc0fe664214ea6e7e27cf1bd6bf124b75379022"} Oct 02 07:35:52 crc kubenswrapper[4829]: I1002 07:35:52.302140 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerStarted","Data":"0ac29cccd1b1d40771e79288e733ace73c59efaf8f51c01cb30363bcc821a605"} Oct 02 07:35:52 crc kubenswrapper[4829]: I1002 07:35:52.533406 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.208948 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.327888 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-httpd-run\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328099 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-logs\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328194 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328431 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-internal-tls-certs\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328491 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shlds\" (UniqueName: \"kubernetes.io/projected/63630a30-0c56-4bdb-b259-984b81733bcc-kube-api-access-shlds\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328532 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-config-data\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328579 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-combined-ca-bundle\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.328618 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-scripts\") pod \"63630a30-0c56-4bdb-b259-984b81733bcc\" (UID: \"63630a30-0c56-4bdb-b259-984b81733bcc\") " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.344970 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-logs" (OuterVolumeSpecName: "logs") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.347077 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.351900 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-scripts" (OuterVolumeSpecName: "scripts") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.355662 4829 generic.go:334] "Generic (PLEG): container finished" podID="63630a30-0c56-4bdb-b259-984b81733bcc" containerID="6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62" exitCode=0 Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.355743 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63630a30-0c56-4bdb-b259-984b81733bcc","Type":"ContainerDied","Data":"6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62"} Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.355768 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.355791 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"63630a30-0c56-4bdb-b259-984b81733bcc","Type":"ContainerDied","Data":"951bbf7fcadc6852f1e4a91df2a509db3739c51d2b1ebfec874802ff3786f3dc"} Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.355812 4829 scope.go:117] "RemoveContainer" containerID="6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.359356 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.361788 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f017e85-e428-48e7-9b00-705ca534019d","Type":"ContainerStarted","Data":"a2efe9eb5ea4f804ec56f47b17e86ac22c20e382cce97b40615df1b9aa0637cc"} Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.361827 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f017e85-e428-48e7-9b00-705ca534019d","Type":"ContainerStarted","Data":"b00ad41957160dc46e092ab6bb10ec7cc892f89f0f3d5c2b6ee0f4348f208cf5"} Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.365477 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63630a30-0c56-4bdb-b259-984b81733bcc-kube-api-access-shlds" (OuterVolumeSpecName: "kube-api-access-shlds") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "kube-api-access-shlds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.392241 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.403411 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.406662 4829 scope.go:117] "RemoveContainer" containerID="41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.410104 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerStarted","Data":"10c15cf80db82ba5bd30c90d96fc82e96323d1446fe73ab8e82b4efcabea796d"} Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.412305 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-config-data" (OuterVolumeSpecName: "config-data") pod "63630a30-0c56-4bdb-b259-984b81733bcc" (UID: "63630a30-0c56-4bdb-b259-984b81733bcc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441426 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441490 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441504 4829 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441518 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shlds\" (UniqueName: \"kubernetes.io/projected/63630a30-0c56-4bdb-b259-984b81733bcc-kube-api-access-shlds\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441532 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441544 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441554 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/63630a30-0c56-4bdb-b259-984b81733bcc-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.441564 4829 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/63630a30-0c56-4bdb-b259-984b81733bcc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.453507 4829 scope.go:117] "RemoveContainer" containerID="6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62" Oct 02 07:35:53 crc kubenswrapper[4829]: E1002 07:35:53.453934 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62\": container with ID starting with 6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62 not found: ID does not exist" containerID="6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.453988 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62"} err="failed to get container status \"6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62\": rpc error: code = NotFound desc = could not find container \"6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62\": container with ID starting with 6c3b5611eadcb473b9d4c245123b987b36f22037e42133188efcf52274affa62 not found: ID does not exist" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.454026 4829 scope.go:117] "RemoveContainer" containerID="41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b" Oct 02 07:35:53 crc kubenswrapper[4829]: E1002 07:35:53.454395 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b\": container with ID starting with 41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b not found: ID does not exist" containerID="41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.454437 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b"} err="failed to get container status \"41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b\": rpc error: code = NotFound desc = could not find container \"41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b\": container with ID starting with 41085bd43663f101e6b2e7e67c64f4f7fc7f884b75188cdd6c0cfec7def74e9b not found: ID does not exist" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.471864 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.473219 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8104fa43-766f-4747-bfde-8efb050e4629" path="/var/lib/kubelet/pods/8104fa43-766f-4747-bfde-8efb050e4629/volumes" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.543359 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.681430 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.701549 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.716151 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:35:53 crc kubenswrapper[4829]: E1002 07:35:53.716602 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-log" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.716617 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-log" Oct 02 07:35:53 crc kubenswrapper[4829]: E1002 07:35:53.716648 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-httpd" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.716656 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-httpd" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.716839 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-httpd" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.716865 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" containerName="glance-log" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.717859 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.724386 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.727700 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.749634 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.849132 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.849420 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-logs\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.849542 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.849688 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.849824 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.849974 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.850127 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.850282 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvvxf\" (UniqueName: \"kubernetes.io/projected/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-kube-api-access-jvvxf\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.951700 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952066 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvvxf\" (UniqueName: \"kubernetes.io/projected/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-kube-api-access-jvvxf\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952178 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952214 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-logs\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952271 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952297 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952341 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952377 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.952807 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.953475 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.957474 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-logs\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.958304 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.959216 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.961639 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.962076 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.968686 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvvxf\" (UniqueName: \"kubernetes.io/projected/9c3dbee2-4ec4-4d14-8e07-8a04428f0429-kube-api-access-jvvxf\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:53 crc kubenswrapper[4829]: I1002 07:35:53.986627 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"9c3dbee2-4ec4-4d14-8e07-8a04428f0429\") " pod="openstack/glance-default-internal-api-0" Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.051147 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.421081 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3f017e85-e428-48e7-9b00-705ca534019d","Type":"ContainerStarted","Data":"9dee58c876a8155b127d147079ac4910c6a7461521bea83a77c52bf64b9d2557"} Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.425100 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerStarted","Data":"e0620f7d80c4991366ae665e08ffe53fe3c2088a8b986dc971a30d205256ed27"} Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.425534 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.425558 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="proxy-httpd" containerID="cri-o://e0620f7d80c4991366ae665e08ffe53fe3c2088a8b986dc971a30d205256ed27" gracePeriod=30 Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.425564 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="sg-core" containerID="cri-o://10c15cf80db82ba5bd30c90d96fc82e96323d1446fe73ab8e82b4efcabea796d" gracePeriod=30 Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.425645 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-notification-agent" containerID="cri-o://2f99639131442bb6a98dc44bcbc0fe664214ea6e7e27cf1bd6bf124b75379022" gracePeriod=30 Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.425789 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-central-agent" containerID="cri-o://0ac29cccd1b1d40771e79288e733ace73c59efaf8f51c01cb30363bcc821a605" gracePeriod=30 Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.448869 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.448848922 podStartE2EDuration="3.448848922s" podCreationTimestamp="2025-10-02 07:35:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:54.441904278 +0000 UTC m=+1145.781552683" watchObservedRunningTime="2025-10-02 07:35:54.448848922 +0000 UTC m=+1145.788497327" Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.618859 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.109937285 podStartE2EDuration="5.618841537s" podCreationTimestamp="2025-10-02 07:35:49 +0000 UTC" firstStartedPulling="2025-10-02 07:35:50.383058701 +0000 UTC m=+1141.722707106" lastFinishedPulling="2025-10-02 07:35:53.891962953 +0000 UTC m=+1145.231611358" observedRunningTime="2025-10-02 07:35:54.463101942 +0000 UTC m=+1145.802750347" watchObservedRunningTime="2025-10-02 07:35:54.618841537 +0000 UTC m=+1145.958489942" Oct 02 07:35:54 crc kubenswrapper[4829]: I1002 07:35:54.621978 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 07:35:54 crc kubenswrapper[4829]: W1002 07:35:54.629513 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c3dbee2_4ec4_4d14_8e07_8a04428f0429.slice/crio-ecbb532a6b1731f8f37c4471f3aefee4b921ea7ffcb4a773e20af14cd5b7b6cc WatchSource:0}: Error finding container ecbb532a6b1731f8f37c4471f3aefee4b921ea7ffcb4a773e20af14cd5b7b6cc: Status 404 returned error can't find the container with id ecbb532a6b1731f8f37c4471f3aefee4b921ea7ffcb4a773e20af14cd5b7b6cc Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.329011 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.329351 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.329543 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.330377 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae2b08236229e74b63b3e0438daf20903c2e9ff5c5e97936f06ade85ab34a7a5"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.330444 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://ae2b08236229e74b63b3e0438daf20903c2e9ff5c5e97936f06ade85ab34a7a5" gracePeriod=600 Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.439373 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9c3dbee2-4ec4-4d14-8e07-8a04428f0429","Type":"ContainerStarted","Data":"886cbacc4092bb437e8b0b06b85b3e46e11259da362a9d6873446d435ce0f03d"} Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.439414 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9c3dbee2-4ec4-4d14-8e07-8a04428f0429","Type":"ContainerStarted","Data":"ecbb532a6b1731f8f37c4471f3aefee4b921ea7ffcb4a773e20af14cd5b7b6cc"} Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.443968 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerID="e0620f7d80c4991366ae665e08ffe53fe3c2088a8b986dc971a30d205256ed27" exitCode=0 Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444008 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerID="10c15cf80db82ba5bd30c90d96fc82e96323d1446fe73ab8e82b4efcabea796d" exitCode=2 Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444035 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerID="2f99639131442bb6a98dc44bcbc0fe664214ea6e7e27cf1bd6bf124b75379022" exitCode=0 Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444044 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerID="0ac29cccd1b1d40771e79288e733ace73c59efaf8f51c01cb30363bcc821a605" exitCode=0 Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444165 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerDied","Data":"e0620f7d80c4991366ae665e08ffe53fe3c2088a8b986dc971a30d205256ed27"} Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444211 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerDied","Data":"10c15cf80db82ba5bd30c90d96fc82e96323d1446fe73ab8e82b4efcabea796d"} Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444246 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerDied","Data":"2f99639131442bb6a98dc44bcbc0fe664214ea6e7e27cf1bd6bf124b75379022"} Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.444256 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerDied","Data":"0ac29cccd1b1d40771e79288e733ace73c59efaf8f51c01cb30363bcc821a605"} Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.476811 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63630a30-0c56-4bdb-b259-984b81733bcc" path="/var/lib/kubelet/pods/63630a30-0c56-4bdb-b259-984b81733bcc/volumes" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.744874 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.884865 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-log-httpd\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.884930 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-scripts\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885027 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-sg-core-conf-yaml\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885059 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-run-httpd\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885179 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-config-data\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885288 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfpwx\" (UniqueName: \"kubernetes.io/projected/4ba3b403-2a6c-4db1-add6-010fbd41728b-kube-api-access-qfpwx\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885328 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-combined-ca-bundle\") pod \"4ba3b403-2a6c-4db1-add6-010fbd41728b\" (UID: \"4ba3b403-2a6c-4db1-add6-010fbd41728b\") " Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885500 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885810 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.885962 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.894720 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba3b403-2a6c-4db1-add6-010fbd41728b-kube-api-access-qfpwx" (OuterVolumeSpecName: "kube-api-access-qfpwx") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "kube-api-access-qfpwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.894727 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-scripts" (OuterVolumeSpecName: "scripts") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.926434 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.985191 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.987322 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.987355 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4ba3b403-2a6c-4db1-add6-010fbd41728b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.987366 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfpwx\" (UniqueName: \"kubernetes.io/projected/4ba3b403-2a6c-4db1-add6-010fbd41728b-kube-api-access-qfpwx\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.987378 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:55 crc kubenswrapper[4829]: I1002 07:35:55.987386 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.003409 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-config-data" (OuterVolumeSpecName: "config-data") pod "4ba3b403-2a6c-4db1-add6-010fbd41728b" (UID: "4ba3b403-2a6c-4db1-add6-010fbd41728b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.089539 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ba3b403-2a6c-4db1-add6-010fbd41728b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.457576 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="ae2b08236229e74b63b3e0438daf20903c2e9ff5c5e97936f06ade85ab34a7a5" exitCode=0 Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.457675 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"ae2b08236229e74b63b3e0438daf20903c2e9ff5c5e97936f06ade85ab34a7a5"} Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.457734 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"17a95fceb29822d3307537c1facf40f10ec266fc571083711123b45f3e6f1880"} Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.457756 4829 scope.go:117] "RemoveContainer" containerID="34b90026201c24a6d0bc1440495bb6980e11f5b356c27ed499750d010a61ec35" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.461051 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9c3dbee2-4ec4-4d14-8e07-8a04428f0429","Type":"ContainerStarted","Data":"1111d4ad26e8894f9cf5b2f352629063d55d3f101e23f38a3968039875aa0861"} Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.469263 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4ba3b403-2a6c-4db1-add6-010fbd41728b","Type":"ContainerDied","Data":"429ab66419b3f97cf2f9fdca2437d8575f4b1f168516a74278dd206482f45a48"} Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.469302 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.503712 4829 scope.go:117] "RemoveContainer" containerID="e0620f7d80c4991366ae665e08ffe53fe3c2088a8b986dc971a30d205256ed27" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.510140 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.510124753 podStartE2EDuration="3.510124753s" podCreationTimestamp="2025-10-02 07:35:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:35:56.507356274 +0000 UTC m=+1147.847004679" watchObservedRunningTime="2025-10-02 07:35:56.510124753 +0000 UTC m=+1147.849773158" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.531848 4829 scope.go:117] "RemoveContainer" containerID="10c15cf80db82ba5bd30c90d96fc82e96323d1446fe73ab8e82b4efcabea796d" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.536435 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.544286 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.562082 4829 scope.go:117] "RemoveContainer" containerID="2f99639131442bb6a98dc44bcbc0fe664214ea6e7e27cf1bd6bf124b75379022" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.563539 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:56 crc kubenswrapper[4829]: E1002 07:35:56.563929 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-notification-agent" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.563946 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-notification-agent" Oct 02 07:35:56 crc kubenswrapper[4829]: E1002 07:35:56.563971 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="proxy-httpd" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.563980 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="proxy-httpd" Oct 02 07:35:56 crc kubenswrapper[4829]: E1002 07:35:56.563995 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="sg-core" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.564001 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="sg-core" Oct 02 07:35:56 crc kubenswrapper[4829]: E1002 07:35:56.564012 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-central-agent" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.564017 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-central-agent" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.564186 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="proxy-httpd" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.564200 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-central-agent" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.564214 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="ceilometer-notification-agent" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.569117 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" containerName="sg-core" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.571646 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.575916 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.576067 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.582472 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.589400 4829 scope.go:117] "RemoveContainer" containerID="0ac29cccd1b1d40771e79288e733ace73c59efaf8f51c01cb30363bcc821a605" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.699719 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.699797 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-run-httpd\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.699880 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-scripts\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.699985 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.700194 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-log-httpd\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.700430 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkdf7\" (UniqueName: \"kubernetes.io/projected/a1606fa8-8934-49e1-89be-ea2948f7e740-kube-api-access-bkdf7\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.700475 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-config-data\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802281 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkdf7\" (UniqueName: \"kubernetes.io/projected/a1606fa8-8934-49e1-89be-ea2948f7e740-kube-api-access-bkdf7\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802324 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-config-data\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802400 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802440 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-run-httpd\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802490 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-scripts\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802512 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.802561 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-log-httpd\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.803222 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-run-httpd\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.803291 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-log-httpd\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.810794 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.810913 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-scripts\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.811175 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.811381 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-config-data\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.825567 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkdf7\" (UniqueName: \"kubernetes.io/projected/a1606fa8-8934-49e1-89be-ea2948f7e740-kube-api-access-bkdf7\") pod \"ceilometer-0\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " pod="openstack/ceilometer-0" Oct 02 07:35:56 crc kubenswrapper[4829]: I1002 07:35:56.887805 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:35:57 crc kubenswrapper[4829]: I1002 07:35:57.378699 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:35:57 crc kubenswrapper[4829]: W1002 07:35:57.382058 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1606fa8_8934_49e1_89be_ea2948f7e740.slice/crio-39a0b27eaedb3ad6834747dcd62865a47d3be32b05f7618f14514e927f5e1d80 WatchSource:0}: Error finding container 39a0b27eaedb3ad6834747dcd62865a47d3be32b05f7618f14514e927f5e1d80: Status 404 returned error can't find the container with id 39a0b27eaedb3ad6834747dcd62865a47d3be32b05f7618f14514e927f5e1d80 Oct 02 07:35:57 crc kubenswrapper[4829]: I1002 07:35:57.480518 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba3b403-2a6c-4db1-add6-010fbd41728b" path="/var/lib/kubelet/pods/4ba3b403-2a6c-4db1-add6-010fbd41728b/volumes" Oct 02 07:35:57 crc kubenswrapper[4829]: I1002 07:35:57.501274 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerStarted","Data":"39a0b27eaedb3ad6834747dcd62865a47d3be32b05f7618f14514e927f5e1d80"} Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.512518 4829 generic.go:334] "Generic (PLEG): container finished" podID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerID="9814c7fd97f5c331f361eb5c9bf44206b3cba730dc3c445f6f3835986ffb7bef" exitCode=137 Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.512596 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dcfb8bd44-ws7cq" event={"ID":"3a176a27-612f-499a-82c4-645d9cdd0b8b","Type":"ContainerDied","Data":"9814c7fd97f5c331f361eb5c9bf44206b3cba730dc3c445f6f3835986ffb7bef"} Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.513025 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7dcfb8bd44-ws7cq" event={"ID":"3a176a27-612f-499a-82c4-645d9cdd0b8b","Type":"ContainerDied","Data":"b67312416317da230b283b4ee1b885cccef967d894ccafcc41f72dc87b5d9b0c"} Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.513041 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b67312416317da230b283b4ee1b885cccef967d894ccafcc41f72dc87b5d9b0c" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.514622 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerStarted","Data":"8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb"} Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.595078 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642344 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-secret-key\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642435 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a176a27-612f-499a-82c4-645d9cdd0b8b-logs\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642473 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-config-data\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642537 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlxn6\" (UniqueName: \"kubernetes.io/projected/3a176a27-612f-499a-82c4-645d9cdd0b8b-kube-api-access-jlxn6\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642598 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-combined-ca-bundle\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642621 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-scripts\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.642714 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-tls-certs\") pod \"3a176a27-612f-499a-82c4-645d9cdd0b8b\" (UID: \"3a176a27-612f-499a-82c4-645d9cdd0b8b\") " Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.643022 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a176a27-612f-499a-82c4-645d9cdd0b8b-logs" (OuterVolumeSpecName: "logs") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.650562 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a176a27-612f-499a-82c4-645d9cdd0b8b-kube-api-access-jlxn6" (OuterVolumeSpecName: "kube-api-access-jlxn6") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "kube-api-access-jlxn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.651116 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.693737 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-config-data" (OuterVolumeSpecName: "config-data") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.701062 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-scripts" (OuterVolumeSpecName: "scripts") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.735094 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.744727 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.744754 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.744763 4829 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.744774 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a176a27-612f-499a-82c4-645d9cdd0b8b-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.744782 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3a176a27-612f-499a-82c4-645d9cdd0b8b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.744790 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlxn6\" (UniqueName: \"kubernetes.io/projected/3a176a27-612f-499a-82c4-645d9cdd0b8b-kube-api-access-jlxn6\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.760538 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3a176a27-612f-499a-82c4-645d9cdd0b8b" (UID: "3a176a27-612f-499a-82c4-645d9cdd0b8b"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:35:58 crc kubenswrapper[4829]: I1002 07:35:58.846750 4829 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a176a27-612f-499a-82c4-645d9cdd0b8b-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:35:59 crc kubenswrapper[4829]: I1002 07:35:59.546416 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7dcfb8bd44-ws7cq" Oct 02 07:35:59 crc kubenswrapper[4829]: I1002 07:35:59.547289 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerStarted","Data":"feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd"} Oct 02 07:35:59 crc kubenswrapper[4829]: I1002 07:35:59.547339 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerStarted","Data":"88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad"} Oct 02 07:35:59 crc kubenswrapper[4829]: I1002 07:35:59.571609 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7dcfb8bd44-ws7cq"] Oct 02 07:35:59 crc kubenswrapper[4829]: I1002 07:35:59.582697 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7dcfb8bd44-ws7cq"] Oct 02 07:36:00 crc kubenswrapper[4829]: I1002 07:36:00.297677 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.471133 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" path="/var/lib/kubelet/pods/3a176a27-612f-499a-82c4-645d9cdd0b8b/volumes" Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.567890 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerStarted","Data":"25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb"} Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.568307 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.568122 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="proxy-httpd" containerID="cri-o://25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb" gracePeriod=30 Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.568085 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-central-agent" containerID="cri-o://8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb" gracePeriod=30 Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.568169 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="sg-core" containerID="cri-o://feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd" gracePeriod=30 Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.568179 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-notification-agent" containerID="cri-o://88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad" gracePeriod=30 Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.593464 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.563070485 podStartE2EDuration="5.593445047s" podCreationTimestamp="2025-10-02 07:35:56 +0000 UTC" firstStartedPulling="2025-10-02 07:35:57.385626243 +0000 UTC m=+1148.725274648" lastFinishedPulling="2025-10-02 07:36:00.416000805 +0000 UTC m=+1151.755649210" observedRunningTime="2025-10-02 07:36:01.589469239 +0000 UTC m=+1152.929117644" watchObservedRunningTime="2025-10-02 07:36:01.593445047 +0000 UTC m=+1152.933093452" Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.946289 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.946562 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 07:36:01 crc kubenswrapper[4829]: I1002 07:36:01.976788 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.003690 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.586296 4829 generic.go:334] "Generic (PLEG): container finished" podID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerID="25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb" exitCode=0 Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.587138 4829 generic.go:334] "Generic (PLEG): container finished" podID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerID="feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd" exitCode=2 Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.587207 4829 generic.go:334] "Generic (PLEG): container finished" podID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerID="88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad" exitCode=0 Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.586499 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerDied","Data":"25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb"} Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.588403 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.588441 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerDied","Data":"feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd"} Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.588473 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 07:36:02 crc kubenswrapper[4829]: I1002 07:36:02.588492 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerDied","Data":"88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad"} Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.477687 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.550958 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-config-data\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.551398 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-run-httpd\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.551440 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkdf7\" (UniqueName: \"kubernetes.io/projected/a1606fa8-8934-49e1-89be-ea2948f7e740-kube-api-access-bkdf7\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.551464 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-log-httpd\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.551508 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-scripts\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.551534 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-sg-core-conf-yaml\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.551668 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-combined-ca-bundle\") pod \"a1606fa8-8934-49e1-89be-ea2948f7e740\" (UID: \"a1606fa8-8934-49e1-89be-ea2948f7e740\") " Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.552041 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.552511 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.557594 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.558504 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-scripts" (OuterVolumeSpecName: "scripts") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.564463 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1606fa8-8934-49e1-89be-ea2948f7e740-kube-api-access-bkdf7" (OuterVolumeSpecName: "kube-api-access-bkdf7") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "kube-api-access-bkdf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.586446 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.597713 4829 generic.go:334] "Generic (PLEG): container finished" podID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerID="8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb" exitCode=0 Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.597817 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.597838 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerDied","Data":"8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb"} Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.598165 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a1606fa8-8934-49e1-89be-ea2948f7e740","Type":"ContainerDied","Data":"39a0b27eaedb3ad6834747dcd62865a47d3be32b05f7618f14514e927f5e1d80"} Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.598247 4829 scope.go:117] "RemoveContainer" containerID="25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.654676 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a1606fa8-8934-49e1-89be-ea2948f7e740-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.654705 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkdf7\" (UniqueName: \"kubernetes.io/projected/a1606fa8-8934-49e1-89be-ea2948f7e740-kube-api-access-bkdf7\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.654719 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.654731 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.655355 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.661213 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-config-data" (OuterVolumeSpecName: "config-data") pod "a1606fa8-8934-49e1-89be-ea2948f7e740" (UID: "a1606fa8-8934-49e1-89be-ea2948f7e740"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.673874 4829 scope.go:117] "RemoveContainer" containerID="feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.704631 4829 scope.go:117] "RemoveContainer" containerID="88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.726272 4829 scope.go:117] "RemoveContainer" containerID="8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.744319 4829 scope.go:117] "RemoveContainer" containerID="25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.744720 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb\": container with ID starting with 25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb not found: ID does not exist" containerID="25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.744756 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb"} err="failed to get container status \"25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb\": rpc error: code = NotFound desc = could not find container \"25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb\": container with ID starting with 25fac82fee0ab75be160e71838344e1d3006b7c614755a637d94ea49948439fb not found: ID does not exist" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.744789 4829 scope.go:117] "RemoveContainer" containerID="feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.745098 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd\": container with ID starting with feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd not found: ID does not exist" containerID="feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.745123 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd"} err="failed to get container status \"feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd\": rpc error: code = NotFound desc = could not find container \"feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd\": container with ID starting with feeb4b26e25909dbe0c845b24fc3b81640599c02dc57e911faf49e5f929255bd not found: ID does not exist" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.745136 4829 scope.go:117] "RemoveContainer" containerID="88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.745420 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad\": container with ID starting with 88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad not found: ID does not exist" containerID="88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.745465 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad"} err="failed to get container status \"88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad\": rpc error: code = NotFound desc = could not find container \"88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad\": container with ID starting with 88fef11653d32bdcf7076b062f1263d8d1c8bd48e165b66588dab797f5c924ad not found: ID does not exist" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.745495 4829 scope.go:117] "RemoveContainer" containerID="8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.745865 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb\": container with ID starting with 8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb not found: ID does not exist" containerID="8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.745897 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb"} err="failed to get container status \"8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb\": rpc error: code = NotFound desc = could not find container \"8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb\": container with ID starting with 8b801fd6508f45608de72a16e1363f3e9b1397869483afad5670a677615202cb not found: ID does not exist" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.756501 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.756537 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1606fa8-8934-49e1-89be-ea2948f7e740-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.937135 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.947279 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966159 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.966667 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="sg-core" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966696 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="sg-core" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.966732 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-central-agent" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966772 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-central-agent" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.966787 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966795 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.966806 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon-log" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966837 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon-log" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.966865 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-notification-agent" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966873 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-notification-agent" Oct 02 07:36:03 crc kubenswrapper[4829]: E1002 07:36:03.966896 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="proxy-httpd" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.966904 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="proxy-httpd" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.967461 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.967485 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-central-agent" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.967494 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a176a27-612f-499a-82c4-645d9cdd0b8b" containerName="horizon-log" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.967509 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="ceilometer-notification-agent" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.967526 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="sg-core" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.967544 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" containerName="proxy-httpd" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.969737 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.976739 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.976962 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:36:03 crc kubenswrapper[4829]: I1002 07:36:03.987304 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.052690 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.054178 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061698 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-config-data\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061747 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrfvn\" (UniqueName: \"kubernetes.io/projected/4c960b00-3055-4155-8ff2-84bfc19182f8-kube-api-access-mrfvn\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061800 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061816 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-scripts\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061853 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061906 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-run-httpd\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.061925 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-log-httpd\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.083799 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.099226 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.163447 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-run-httpd\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.163991 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-log-httpd\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.164181 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-config-data\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.164368 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrfvn\" (UniqueName: \"kubernetes.io/projected/4c960b00-3055-4155-8ff2-84bfc19182f8-kube-api-access-mrfvn\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.164513 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.164647 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-scripts\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.164829 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.165917 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-log-httpd\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.166170 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-run-httpd\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.170824 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.171162 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-scripts\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.172060 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-config-data\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.173850 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.181985 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrfvn\" (UniqueName: \"kubernetes.io/projected/4c960b00-3055-4155-8ff2-84bfc19182f8-kube-api-access-mrfvn\") pod \"ceilometer-0\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.301491 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.506209 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.604630 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0c73-account-create-648bq"] Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.605916 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.615149 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0c73-account-create-648bq"] Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.616544 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.616656 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.628073 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.628108 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.784608 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpczj\" (UniqueName: \"kubernetes.io/projected/19037609-49d9-4e96-b537-30e2765c6f26-kube-api-access-qpczj\") pod \"nova-api-0c73-account-create-648bq\" (UID: \"19037609-49d9-4e96-b537-30e2765c6f26\") " pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.790152 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.812365 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-faf0-account-create-qxxsr"] Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.813934 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.818074 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.825476 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-faf0-account-create-qxxsr"] Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.886645 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpczj\" (UniqueName: \"kubernetes.io/projected/19037609-49d9-4e96-b537-30e2765c6f26-kube-api-access-qpczj\") pod \"nova-api-0c73-account-create-648bq\" (UID: \"19037609-49d9-4e96-b537-30e2765c6f26\") " pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.908154 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpczj\" (UniqueName: \"kubernetes.io/projected/19037609-49d9-4e96-b537-30e2765c6f26-kube-api-access-qpczj\") pod \"nova-api-0c73-account-create-648bq\" (UID: \"19037609-49d9-4e96-b537-30e2765c6f26\") " pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.936134 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:04 crc kubenswrapper[4829]: I1002 07:36:04.988599 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb9md\" (UniqueName: \"kubernetes.io/projected/9460237e-d366-4526-a859-06daeae67fce-kube-api-access-qb9md\") pod \"nova-cell0-faf0-account-create-qxxsr\" (UID: \"9460237e-d366-4526-a859-06daeae67fce\") " pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.032485 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-3b7f-account-create-dpq82"] Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.034064 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.036965 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.053107 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3b7f-account-create-dpq82"] Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.090274 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb9md\" (UniqueName: \"kubernetes.io/projected/9460237e-d366-4526-a859-06daeae67fce-kube-api-access-qb9md\") pod \"nova-cell0-faf0-account-create-qxxsr\" (UID: \"9460237e-d366-4526-a859-06daeae67fce\") " pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.108278 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb9md\" (UniqueName: \"kubernetes.io/projected/9460237e-d366-4526-a859-06daeae67fce-kube-api-access-qb9md\") pod \"nova-cell0-faf0-account-create-qxxsr\" (UID: \"9460237e-d366-4526-a859-06daeae67fce\") " pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.134938 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.193373 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdm9j\" (UniqueName: \"kubernetes.io/projected/ecce1ed1-3a3e-4576-853e-0972201544e9-kube-api-access-xdm9j\") pod \"nova-cell1-3b7f-account-create-dpq82\" (UID: \"ecce1ed1-3a3e-4576-853e-0972201544e9\") " pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.294877 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdm9j\" (UniqueName: \"kubernetes.io/projected/ecce1ed1-3a3e-4576-853e-0972201544e9-kube-api-access-xdm9j\") pod \"nova-cell1-3b7f-account-create-dpq82\" (UID: \"ecce1ed1-3a3e-4576-853e-0972201544e9\") " pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.324058 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdm9j\" (UniqueName: \"kubernetes.io/projected/ecce1ed1-3a3e-4576-853e-0972201544e9-kube-api-access-xdm9j\") pod \"nova-cell1-3b7f-account-create-dpq82\" (UID: \"ecce1ed1-3a3e-4576-853e-0972201544e9\") " pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.393522 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.487583 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1606fa8-8934-49e1-89be-ea2948f7e740" path="/var/lib/kubelet/pods/a1606fa8-8934-49e1-89be-ea2948f7e740/volumes" Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.492023 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0c73-account-create-648bq"] Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.671796 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-faf0-account-create-qxxsr"] Oct 02 07:36:05 crc kubenswrapper[4829]: W1002 07:36:05.683364 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9460237e_d366_4526_a859_06daeae67fce.slice/crio-72c950f8fd14439bf3cf546f7e534aff112e5fbe8ce28c5924f384e31a8f2eb7 WatchSource:0}: Error finding container 72c950f8fd14439bf3cf546f7e534aff112e5fbe8ce28c5924f384e31a8f2eb7: Status 404 returned error can't find the container with id 72c950f8fd14439bf3cf546f7e534aff112e5fbe8ce28c5924f384e31a8f2eb7 Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.683617 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerStarted","Data":"dbc0eef633782b06627a5a99bc93afe8c46fe74d53fb35e84fff202334b6806b"} Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.683655 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerStarted","Data":"bc7a8f855d3d9f665e0ec9a869d46c6d0fc0b6480b284c9b3687e824c5a07e84"} Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.696199 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0c73-account-create-648bq" event={"ID":"19037609-49d9-4e96-b537-30e2765c6f26","Type":"ContainerStarted","Data":"7f5d6994081a413b76b768fe14aa02545c642182f428d6a136c0a2e1bc9fcb3f"} Oct 02 07:36:05 crc kubenswrapper[4829]: I1002 07:36:05.855247 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-3b7f-account-create-dpq82"] Oct 02 07:36:05 crc kubenswrapper[4829]: W1002 07:36:05.878327 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecce1ed1_3a3e_4576_853e_0972201544e9.slice/crio-85a352a82f1149bc64f2904d0242d55b8024d6d59836c1491adf1758950cdb5c WatchSource:0}: Error finding container 85a352a82f1149bc64f2904d0242d55b8024d6d59836c1491adf1758950cdb5c: Status 404 returned error can't find the container with id 85a352a82f1149bc64f2904d0242d55b8024d6d59836c1491adf1758950cdb5c Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.703764 4829 generic.go:334] "Generic (PLEG): container finished" podID="ecce1ed1-3a3e-4576-853e-0972201544e9" containerID="fa3755afd0985e5c01dad6075ff0333806be74baa8948041accbb147d52e0b6c" exitCode=0 Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.703861 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3b7f-account-create-dpq82" event={"ID":"ecce1ed1-3a3e-4576-853e-0972201544e9","Type":"ContainerDied","Data":"fa3755afd0985e5c01dad6075ff0333806be74baa8948041accbb147d52e0b6c"} Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.704055 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3b7f-account-create-dpq82" event={"ID":"ecce1ed1-3a3e-4576-853e-0972201544e9","Type":"ContainerStarted","Data":"85a352a82f1149bc64f2904d0242d55b8024d6d59836c1491adf1758950cdb5c"} Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.705943 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerStarted","Data":"c858717621f4511d971b93d5eb4301eb2ae229e56ebf7d7cb53ba56117afb84b"} Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.707591 4829 generic.go:334] "Generic (PLEG): container finished" podID="19037609-49d9-4e96-b537-30e2765c6f26" containerID="c1d874ae94d8d9d7f79e67f5c372a59c2c1c1b604d5709981151cef9e88a3428" exitCode=0 Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.707641 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0c73-account-create-648bq" event={"ID":"19037609-49d9-4e96-b537-30e2765c6f26","Type":"ContainerDied","Data":"c1d874ae94d8d9d7f79e67f5c372a59c2c1c1b604d5709981151cef9e88a3428"} Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.709089 4829 generic.go:334] "Generic (PLEG): container finished" podID="9460237e-d366-4526-a859-06daeae67fce" containerID="4b2aba6614d0ef4b94061fb92b2bd700c14bbc24b3ab2c4572f13109875b58fd" exitCode=0 Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.709149 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-faf0-account-create-qxxsr" event={"ID":"9460237e-d366-4526-a859-06daeae67fce","Type":"ContainerDied","Data":"4b2aba6614d0ef4b94061fb92b2bd700c14bbc24b3ab2c4572f13109875b58fd"} Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.709195 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-faf0-account-create-qxxsr" event={"ID":"9460237e-d366-4526-a859-06daeae67fce","Type":"ContainerStarted","Data":"72c950f8fd14439bf3cf546f7e534aff112e5fbe8ce28c5924f384e31a8f2eb7"} Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.709169 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.709232 4829 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.819632 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:06 crc kubenswrapper[4829]: I1002 07:36:06.967690 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 07:36:07 crc kubenswrapper[4829]: I1002 07:36:07.721841 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerStarted","Data":"78975ac5f0805e63bef5094841aed8683f1304c5947a33cdb4f879e6b40ca006"} Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.378462 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.399085 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.426869 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.471955 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdm9j\" (UniqueName: \"kubernetes.io/projected/ecce1ed1-3a3e-4576-853e-0972201544e9-kube-api-access-xdm9j\") pod \"ecce1ed1-3a3e-4576-853e-0972201544e9\" (UID: \"ecce1ed1-3a3e-4576-853e-0972201544e9\") " Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.472115 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb9md\" (UniqueName: \"kubernetes.io/projected/9460237e-d366-4526-a859-06daeae67fce-kube-api-access-qb9md\") pod \"9460237e-d366-4526-a859-06daeae67fce\" (UID: \"9460237e-d366-4526-a859-06daeae67fce\") " Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.490353 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecce1ed1-3a3e-4576-853e-0972201544e9-kube-api-access-xdm9j" (OuterVolumeSpecName: "kube-api-access-xdm9j") pod "ecce1ed1-3a3e-4576-853e-0972201544e9" (UID: "ecce1ed1-3a3e-4576-853e-0972201544e9"). InnerVolumeSpecName "kube-api-access-xdm9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.490403 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9460237e-d366-4526-a859-06daeae67fce-kube-api-access-qb9md" (OuterVolumeSpecName: "kube-api-access-qb9md") pod "9460237e-d366-4526-a859-06daeae67fce" (UID: "9460237e-d366-4526-a859-06daeae67fce"). InnerVolumeSpecName "kube-api-access-qb9md". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.576914 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpczj\" (UniqueName: \"kubernetes.io/projected/19037609-49d9-4e96-b537-30e2765c6f26-kube-api-access-qpczj\") pod \"19037609-49d9-4e96-b537-30e2765c6f26\" (UID: \"19037609-49d9-4e96-b537-30e2765c6f26\") " Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.577432 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdm9j\" (UniqueName: \"kubernetes.io/projected/ecce1ed1-3a3e-4576-853e-0972201544e9-kube-api-access-xdm9j\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.577449 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb9md\" (UniqueName: \"kubernetes.io/projected/9460237e-d366-4526-a859-06daeae67fce-kube-api-access-qb9md\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.584367 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19037609-49d9-4e96-b537-30e2765c6f26-kube-api-access-qpczj" (OuterVolumeSpecName: "kube-api-access-qpczj") pod "19037609-49d9-4e96-b537-30e2765c6f26" (UID: "19037609-49d9-4e96-b537-30e2765c6f26"). InnerVolumeSpecName "kube-api-access-qpczj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.679492 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpczj\" (UniqueName: \"kubernetes.io/projected/19037609-49d9-4e96-b537-30e2765c6f26-kube-api-access-qpczj\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.736910 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerStarted","Data":"cd92d4b5c7d9add923ce172ba4ecd8cde058338a3d42da3cb7cb459998440dd6"} Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.737015 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.744804 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0c73-account-create-648bq" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.744950 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0c73-account-create-648bq" event={"ID":"19037609-49d9-4e96-b537-30e2765c6f26","Type":"ContainerDied","Data":"7f5d6994081a413b76b768fe14aa02545c642182f428d6a136c0a2e1bc9fcb3f"} Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.744989 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f5d6994081a413b76b768fe14aa02545c642182f428d6a136c0a2e1bc9fcb3f" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.756422 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-faf0-account-create-qxxsr" event={"ID":"9460237e-d366-4526-a859-06daeae67fce","Type":"ContainerDied","Data":"72c950f8fd14439bf3cf546f7e534aff112e5fbe8ce28c5924f384e31a8f2eb7"} Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.756462 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72c950f8fd14439bf3cf546f7e534aff112e5fbe8ce28c5924f384e31a8f2eb7" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.756693 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-faf0-account-create-qxxsr" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.760848 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.273626336 podStartE2EDuration="5.760837328s" podCreationTimestamp="2025-10-02 07:36:03 +0000 UTC" firstStartedPulling="2025-10-02 07:36:04.796588743 +0000 UTC m=+1156.136237148" lastFinishedPulling="2025-10-02 07:36:08.283799725 +0000 UTC m=+1159.623448140" observedRunningTime="2025-10-02 07:36:08.758072919 +0000 UTC m=+1160.097721324" watchObservedRunningTime="2025-10-02 07:36:08.760837328 +0000 UTC m=+1160.100485733" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.764097 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-3b7f-account-create-dpq82" event={"ID":"ecce1ed1-3a3e-4576-853e-0972201544e9","Type":"ContainerDied","Data":"85a352a82f1149bc64f2904d0242d55b8024d6d59836c1491adf1758950cdb5c"} Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.764168 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85a352a82f1149bc64f2904d0242d55b8024d6d59836c1491adf1758950cdb5c" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.764262 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-3b7f-account-create-dpq82" Oct 02 07:36:08 crc kubenswrapper[4829]: I1002 07:36:08.917134 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.107338 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8fgfl"] Oct 02 07:36:10 crc kubenswrapper[4829]: E1002 07:36:10.108075 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9460237e-d366-4526-a859-06daeae67fce" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.108090 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9460237e-d366-4526-a859-06daeae67fce" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: E1002 07:36:10.108135 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecce1ed1-3a3e-4576-853e-0972201544e9" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.108142 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecce1ed1-3a3e-4576-853e-0972201544e9" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: E1002 07:36:10.108150 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19037609-49d9-4e96-b537-30e2765c6f26" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.108157 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="19037609-49d9-4e96-b537-30e2765c6f26" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.108361 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecce1ed1-3a3e-4576-853e-0972201544e9" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.108379 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9460237e-d366-4526-a859-06daeae67fce" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.108392 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="19037609-49d9-4e96-b537-30e2765c6f26" containerName="mariadb-account-create" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.109042 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.113202 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wnqqz" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.113547 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.113670 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.132689 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8fgfl"] Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.206775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.207587 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-scripts\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.207646 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p6vt\" (UniqueName: \"kubernetes.io/projected/f49fa10e-56bf-4af5-874a-9e73967cd27d-kube-api-access-7p6vt\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.207763 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-config-data\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.309200 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p6vt\" (UniqueName: \"kubernetes.io/projected/f49fa10e-56bf-4af5-874a-9e73967cd27d-kube-api-access-7p6vt\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.309260 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-config-data\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.309412 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.309434 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-scripts\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.314762 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-config-data\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.316638 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-scripts\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.323904 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.325551 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p6vt\" (UniqueName: \"kubernetes.io/projected/f49fa10e-56bf-4af5-874a-9e73967cd27d-kube-api-access-7p6vt\") pod \"nova-cell0-conductor-db-sync-8fgfl\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.438896 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.782205 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-central-agent" containerID="cri-o://dbc0eef633782b06627a5a99bc93afe8c46fe74d53fb35e84fff202334b6806b" gracePeriod=30 Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.782657 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="proxy-httpd" containerID="cri-o://cd92d4b5c7d9add923ce172ba4ecd8cde058338a3d42da3cb7cb459998440dd6" gracePeriod=30 Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.782750 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-notification-agent" containerID="cri-o://c858717621f4511d971b93d5eb4301eb2ae229e56ebf7d7cb53ba56117afb84b" gracePeriod=30 Oct 02 07:36:10 crc kubenswrapper[4829]: I1002 07:36:10.782800 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="sg-core" containerID="cri-o://78975ac5f0805e63bef5094841aed8683f1304c5947a33cdb4f879e6b40ca006" gracePeriod=30 Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.032778 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8fgfl"] Oct 02 07:36:11 crc kubenswrapper[4829]: W1002 07:36:11.040563 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf49fa10e_56bf_4af5_874a_9e73967cd27d.slice/crio-acb8efa0395e71a8f3329640e22465e1c61b55b5f234714d56b1c963f75bc7aa WatchSource:0}: Error finding container acb8efa0395e71a8f3329640e22465e1c61b55b5f234714d56b1c963f75bc7aa: Status 404 returned error can't find the container with id acb8efa0395e71a8f3329640e22465e1c61b55b5f234714d56b1c963f75bc7aa Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.802601 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" event={"ID":"f49fa10e-56bf-4af5-874a-9e73967cd27d","Type":"ContainerStarted","Data":"acb8efa0395e71a8f3329640e22465e1c61b55b5f234714d56b1c963f75bc7aa"} Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.806743 4829 generic.go:334] "Generic (PLEG): container finished" podID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerID="cd92d4b5c7d9add923ce172ba4ecd8cde058338a3d42da3cb7cb459998440dd6" exitCode=0 Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.806773 4829 generic.go:334] "Generic (PLEG): container finished" podID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerID="78975ac5f0805e63bef5094841aed8683f1304c5947a33cdb4f879e6b40ca006" exitCode=2 Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.806782 4829 generic.go:334] "Generic (PLEG): container finished" podID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerID="c858717621f4511d971b93d5eb4301eb2ae229e56ebf7d7cb53ba56117afb84b" exitCode=0 Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.806792 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerDied","Data":"cd92d4b5c7d9add923ce172ba4ecd8cde058338a3d42da3cb7cb459998440dd6"} Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.806837 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerDied","Data":"78975ac5f0805e63bef5094841aed8683f1304c5947a33cdb4f879e6b40ca006"} Oct 02 07:36:11 crc kubenswrapper[4829]: I1002 07:36:11.806849 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerDied","Data":"c858717621f4511d971b93d5eb4301eb2ae229e56ebf7d7cb53ba56117afb84b"} Oct 02 07:36:13 crc kubenswrapper[4829]: I1002 07:36:13.852705 4829 generic.go:334] "Generic (PLEG): container finished" podID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerID="dbc0eef633782b06627a5a99bc93afe8c46fe74d53fb35e84fff202334b6806b" exitCode=0 Oct 02 07:36:13 crc kubenswrapper[4829]: I1002 07:36:13.852788 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerDied","Data":"dbc0eef633782b06627a5a99bc93afe8c46fe74d53fb35e84fff202334b6806b"} Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.412087 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462286 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-combined-ca-bundle\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462430 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-scripts\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462513 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-log-httpd\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462649 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-run-httpd\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462677 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-config-data\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462718 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrfvn\" (UniqueName: \"kubernetes.io/projected/4c960b00-3055-4155-8ff2-84bfc19182f8-kube-api-access-mrfvn\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462749 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-sg-core-conf-yaml\") pod \"4c960b00-3055-4155-8ff2-84bfc19182f8\" (UID: \"4c960b00-3055-4155-8ff2-84bfc19182f8\") " Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.462937 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.463036 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.463539 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.463564 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4c960b00-3055-4155-8ff2-84bfc19182f8-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.471665 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-scripts" (OuterVolumeSpecName: "scripts") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.471688 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c960b00-3055-4155-8ff2-84bfc19182f8-kube-api-access-mrfvn" (OuterVolumeSpecName: "kube-api-access-mrfvn") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "kube-api-access-mrfvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.494786 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.556287 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.562480 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-config-data" (OuterVolumeSpecName: "config-data") pod "4c960b00-3055-4155-8ff2-84bfc19182f8" (UID: "4c960b00-3055-4155-8ff2-84bfc19182f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.565011 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrfvn\" (UniqueName: \"kubernetes.io/projected/4c960b00-3055-4155-8ff2-84bfc19182f8-kube-api-access-mrfvn\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.565039 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.565048 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.565058 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.565066 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c960b00-3055-4155-8ff2-84bfc19182f8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.913104 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" event={"ID":"f49fa10e-56bf-4af5-874a-9e73967cd27d","Type":"ContainerStarted","Data":"787dc83cf5df9eefea49e19665d944403ada41509d7f27120d3529a358363a7b"} Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.918916 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4c960b00-3055-4155-8ff2-84bfc19182f8","Type":"ContainerDied","Data":"bc7a8f855d3d9f665e0ec9a869d46c6d0fc0b6480b284c9b3687e824c5a07e84"} Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.918960 4829 scope.go:117] "RemoveContainer" containerID="cd92d4b5c7d9add923ce172ba4ecd8cde058338a3d42da3cb7cb459998440dd6" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.919085 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.943370 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" podStartSLOduration=1.8963143900000001 podStartE2EDuration="8.943348836s" podCreationTimestamp="2025-10-02 07:36:10 +0000 UTC" firstStartedPulling="2025-10-02 07:36:11.042510001 +0000 UTC m=+1162.382158406" lastFinishedPulling="2025-10-02 07:36:18.089544447 +0000 UTC m=+1169.429192852" observedRunningTime="2025-10-02 07:36:18.939139269 +0000 UTC m=+1170.278787714" watchObservedRunningTime="2025-10-02 07:36:18.943348836 +0000 UTC m=+1170.282997241" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.974514 4829 scope.go:117] "RemoveContainer" containerID="78975ac5f0805e63bef5094841aed8683f1304c5947a33cdb4f879e6b40ca006" Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.975469 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:18 crc kubenswrapper[4829]: I1002 07:36:18.984439 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.017676 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:19 crc kubenswrapper[4829]: E1002 07:36:19.018149 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-notification-agent" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018172 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-notification-agent" Oct 02 07:36:19 crc kubenswrapper[4829]: E1002 07:36:19.018185 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="sg-core" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018193 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="sg-core" Oct 02 07:36:19 crc kubenswrapper[4829]: E1002 07:36:19.018243 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-central-agent" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018253 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-central-agent" Oct 02 07:36:19 crc kubenswrapper[4829]: E1002 07:36:19.018269 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="proxy-httpd" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018276 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="proxy-httpd" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018463 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-notification-agent" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018483 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="ceilometer-central-agent" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018494 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="sg-core" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.018516 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" containerName="proxy-httpd" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.021877 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.025543 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.025568 4829 scope.go:117] "RemoveContainer" containerID="c858717621f4511d971b93d5eb4301eb2ae229e56ebf7d7cb53ba56117afb84b" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.025766 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.034079 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.068279 4829 scope.go:117] "RemoveContainer" containerID="dbc0eef633782b06627a5a99bc93afe8c46fe74d53fb35e84fff202334b6806b" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.074750 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.075096 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-config-data\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.075148 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-scripts\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.075181 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-run-httpd\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.075275 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.075312 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjln9\" (UniqueName: \"kubernetes.io/projected/a5d4e255-920f-4f35-89f6-08bc7355ff03-kube-api-access-bjln9\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.075384 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-log-httpd\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.177815 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.177925 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjln9\" (UniqueName: \"kubernetes.io/projected/a5d4e255-920f-4f35-89f6-08bc7355ff03-kube-api-access-bjln9\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.178046 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-log-httpd\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.178186 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.178322 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-config-data\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.178397 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-scripts\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.178449 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-run-httpd\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.178836 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-log-httpd\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.179142 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-run-httpd\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.184109 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-scripts\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.185003 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.187316 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-config-data\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.194735 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.201404 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjln9\" (UniqueName: \"kubernetes.io/projected/a5d4e255-920f-4f35-89f6-08bc7355ff03-kube-api-access-bjln9\") pod \"ceilometer-0\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.368478 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.519189 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c960b00-3055-4155-8ff2-84bfc19182f8" path="/var/lib/kubelet/pods/4c960b00-3055-4155-8ff2-84bfc19182f8/volumes" Oct 02 07:36:19 crc kubenswrapper[4829]: W1002 07:36:19.873024 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5d4e255_920f_4f35_89f6_08bc7355ff03.slice/crio-425329d7716e3b0a3452c719f483c89ff6aad9bd0b818d8e00a8dee558702188 WatchSource:0}: Error finding container 425329d7716e3b0a3452c719f483c89ff6aad9bd0b818d8e00a8dee558702188: Status 404 returned error can't find the container with id 425329d7716e3b0a3452c719f483c89ff6aad9bd0b818d8e00a8dee558702188 Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.875185 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:19 crc kubenswrapper[4829]: I1002 07:36:19.936505 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerStarted","Data":"425329d7716e3b0a3452c719f483c89ff6aad9bd0b818d8e00a8dee558702188"} Oct 02 07:36:20 crc kubenswrapper[4829]: I1002 07:36:20.951068 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerStarted","Data":"13f6ccd385bcbd94cc65b2964ca612bf866091f1e4f8a9526fe7b83dec51ceb4"} Oct 02 07:36:21 crc kubenswrapper[4829]: I1002 07:36:21.966892 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerStarted","Data":"4fbb90c51caee159cef1056dd8484fa71b4577db9ac880c9dc1c443aa3abb4f5"} Oct 02 07:36:21 crc kubenswrapper[4829]: I1002 07:36:21.967376 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerStarted","Data":"b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72"} Oct 02 07:36:23 crc kubenswrapper[4829]: I1002 07:36:23.990209 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerStarted","Data":"541258c680fb40bdf3dfe44364e81d92bf3100c73c2bdbf3cc519b48c3673666"} Oct 02 07:36:23 crc kubenswrapper[4829]: I1002 07:36:23.990913 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:36:29 crc kubenswrapper[4829]: I1002 07:36:29.076052 4829 generic.go:334] "Generic (PLEG): container finished" podID="f49fa10e-56bf-4af5-874a-9e73967cd27d" containerID="787dc83cf5df9eefea49e19665d944403ada41509d7f27120d3529a358363a7b" exitCode=0 Oct 02 07:36:29 crc kubenswrapper[4829]: I1002 07:36:29.076165 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" event={"ID":"f49fa10e-56bf-4af5-874a-9e73967cd27d","Type":"ContainerDied","Data":"787dc83cf5df9eefea49e19665d944403ada41509d7f27120d3529a358363a7b"} Oct 02 07:36:29 crc kubenswrapper[4829]: I1002 07:36:29.106597 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=7.542587456 podStartE2EDuration="11.106576002s" podCreationTimestamp="2025-10-02 07:36:18 +0000 UTC" firstStartedPulling="2025-10-02 07:36:19.875463622 +0000 UTC m=+1171.215112037" lastFinishedPulling="2025-10-02 07:36:23.439452178 +0000 UTC m=+1174.779100583" observedRunningTime="2025-10-02 07:36:24.0178964 +0000 UTC m=+1175.357544805" watchObservedRunningTime="2025-10-02 07:36:29.106576002 +0000 UTC m=+1180.446224407" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.523791 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.613966 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-config-data\") pod \"f49fa10e-56bf-4af5-874a-9e73967cd27d\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.614461 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-combined-ca-bundle\") pod \"f49fa10e-56bf-4af5-874a-9e73967cd27d\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.614505 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-scripts\") pod \"f49fa10e-56bf-4af5-874a-9e73967cd27d\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.614705 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p6vt\" (UniqueName: \"kubernetes.io/projected/f49fa10e-56bf-4af5-874a-9e73967cd27d-kube-api-access-7p6vt\") pod \"f49fa10e-56bf-4af5-874a-9e73967cd27d\" (UID: \"f49fa10e-56bf-4af5-874a-9e73967cd27d\") " Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.620506 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-scripts" (OuterVolumeSpecName: "scripts") pod "f49fa10e-56bf-4af5-874a-9e73967cd27d" (UID: "f49fa10e-56bf-4af5-874a-9e73967cd27d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.627598 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f49fa10e-56bf-4af5-874a-9e73967cd27d-kube-api-access-7p6vt" (OuterVolumeSpecName: "kube-api-access-7p6vt") pod "f49fa10e-56bf-4af5-874a-9e73967cd27d" (UID: "f49fa10e-56bf-4af5-874a-9e73967cd27d"). InnerVolumeSpecName "kube-api-access-7p6vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.653560 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f49fa10e-56bf-4af5-874a-9e73967cd27d" (UID: "f49fa10e-56bf-4af5-874a-9e73967cd27d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.678031 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-config-data" (OuterVolumeSpecName: "config-data") pod "f49fa10e-56bf-4af5-874a-9e73967cd27d" (UID: "f49fa10e-56bf-4af5-874a-9e73967cd27d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.717479 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p6vt\" (UniqueName: \"kubernetes.io/projected/f49fa10e-56bf-4af5-874a-9e73967cd27d-kube-api-access-7p6vt\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.717525 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.717537 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:30 crc kubenswrapper[4829]: I1002 07:36:30.717551 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49fa10e-56bf-4af5-874a-9e73967cd27d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.107729 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" event={"ID":"f49fa10e-56bf-4af5-874a-9e73967cd27d","Type":"ContainerDied","Data":"acb8efa0395e71a8f3329640e22465e1c61b55b5f234714d56b1c963f75bc7aa"} Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.107791 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acb8efa0395e71a8f3329640e22465e1c61b55b5f234714d56b1c963f75bc7aa" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.107842 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-8fgfl" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.271920 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 07:36:31 crc kubenswrapper[4829]: E1002 07:36:31.272667 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f49fa10e-56bf-4af5-874a-9e73967cd27d" containerName="nova-cell0-conductor-db-sync" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.272698 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f49fa10e-56bf-4af5-874a-9e73967cd27d" containerName="nova-cell0-conductor-db-sync" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.273061 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f49fa10e-56bf-4af5-874a-9e73967cd27d" containerName="nova-cell0-conductor-db-sync" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.274398 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.276909 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-wnqqz" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.277608 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.283887 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.437054 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/911b568c-7273-4d79-8213-0d2577e98046-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.437719 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmpgg\" (UniqueName: \"kubernetes.io/projected/911b568c-7273-4d79-8213-0d2577e98046-kube-api-access-lmpgg\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.438375 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b568c-7273-4d79-8213-0d2577e98046-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.540692 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b568c-7273-4d79-8213-0d2577e98046-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.540817 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/911b568c-7273-4d79-8213-0d2577e98046-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.540959 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmpgg\" (UniqueName: \"kubernetes.io/projected/911b568c-7273-4d79-8213-0d2577e98046-kube-api-access-lmpgg\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.548910 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/911b568c-7273-4d79-8213-0d2577e98046-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.555908 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/911b568c-7273-4d79-8213-0d2577e98046-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.567697 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmpgg\" (UniqueName: \"kubernetes.io/projected/911b568c-7273-4d79-8213-0d2577e98046-kube-api-access-lmpgg\") pod \"nova-cell0-conductor-0\" (UID: \"911b568c-7273-4d79-8213-0d2577e98046\") " pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:31 crc kubenswrapper[4829]: I1002 07:36:31.596844 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:32 crc kubenswrapper[4829]: I1002 07:36:32.170168 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 07:36:32 crc kubenswrapper[4829]: W1002 07:36:32.172757 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod911b568c_7273_4d79_8213_0d2577e98046.slice/crio-a4d709d00cd224d307b8e25a4c52345f82fe1a07233d14566826cf649bd7e73b WatchSource:0}: Error finding container a4d709d00cd224d307b8e25a4c52345f82fe1a07233d14566826cf649bd7e73b: Status 404 returned error can't find the container with id a4d709d00cd224d307b8e25a4c52345f82fe1a07233d14566826cf649bd7e73b Oct 02 07:36:33 crc kubenswrapper[4829]: I1002 07:36:33.135888 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"911b568c-7273-4d79-8213-0d2577e98046","Type":"ContainerStarted","Data":"5113fe41b006d3918b97a089dc8d7a4c2f86eb49313e918650e3f96b253e0dfb"} Oct 02 07:36:33 crc kubenswrapper[4829]: I1002 07:36:33.136218 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"911b568c-7273-4d79-8213-0d2577e98046","Type":"ContainerStarted","Data":"a4d709d00cd224d307b8e25a4c52345f82fe1a07233d14566826cf649bd7e73b"} Oct 02 07:36:33 crc kubenswrapper[4829]: I1002 07:36:33.136731 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:33 crc kubenswrapper[4829]: I1002 07:36:33.167476 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.167457666 podStartE2EDuration="2.167457666s" podCreationTimestamp="2025-10-02 07:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:33.165882305 +0000 UTC m=+1184.505530750" watchObservedRunningTime="2025-10-02 07:36:33.167457666 +0000 UTC m=+1184.507106081" Oct 02 07:36:41 crc kubenswrapper[4829]: I1002 07:36:41.628946 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.253885 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-4jqqc"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.255286 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.257983 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.260559 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.267991 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4jqqc"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.338455 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.338513 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-config-data\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.338648 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-scripts\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.338705 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr9xl\" (UniqueName: \"kubernetes.io/projected/0f195e99-4316-4030-9e49-1bf368defe41-kube-api-access-vr9xl\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.380646 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.382108 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.390499 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.397957 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440679 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440731 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-config-data\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440777 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prtv9\" (UniqueName: \"kubernetes.io/projected/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-kube-api-access-prtv9\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440840 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-scripts\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440860 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr9xl\" (UniqueName: \"kubernetes.io/projected/0f195e99-4316-4030-9e49-1bf368defe41-kube-api-access-vr9xl\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440915 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-logs\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440965 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.440984 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-config-data\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.448903 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-scripts\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.449394 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.451166 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-config-data\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.473740 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr9xl\" (UniqueName: \"kubernetes.io/projected/0f195e99-4316-4030-9e49-1bf368defe41-kube-api-access-vr9xl\") pod \"nova-cell0-cell-mapping-4jqqc\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.498306 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.499852 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.502909 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.505942 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.542791 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-logs\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.542871 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.542892 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-config-data\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.542950 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prtv9\" (UniqueName: \"kubernetes.io/projected/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-kube-api-access-prtv9\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.543619 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-logs\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.560642 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.562215 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.570015 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.574908 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.588901 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-config-data\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.589343 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.592939 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prtv9\" (UniqueName: \"kubernetes.io/projected/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-kube-api-access-prtv9\") pod \"nova-api-0\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.632507 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.647668 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnfth\" (UniqueName: \"kubernetes.io/projected/958f85b0-6cda-4f95-b713-4b0b59237d24-kube-api-access-qnfth\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.648045 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-config-data\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.648131 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.690104 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.691368 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.693319 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.706978 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.719377 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.724002 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gplfb"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.725968 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.738185 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gplfb"] Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749413 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnfth\" (UniqueName: \"kubernetes.io/projected/958f85b0-6cda-4f95-b713-4b0b59237d24-kube-api-access-qnfth\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749489 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749509 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-config-data\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749528 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prtdj\" (UniqueName: \"kubernetes.io/projected/9c6f6e08-5872-47ea-86d3-a185f90d981d-kube-api-access-prtdj\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749545 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6f6e08-5872-47ea-86d3-a185f90d981d-logs\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749564 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.749615 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-config-data\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.756439 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-config-data\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.756646 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.779088 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnfth\" (UniqueName: \"kubernetes.io/projected/958f85b0-6cda-4f95-b713-4b0b59237d24-kube-api-access-qnfth\") pod \"nova-scheduler-0\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " pod="openstack/nova-scheduler-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.850938 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851327 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b28nk\" (UniqueName: \"kubernetes.io/projected/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-kube-api-access-b28nk\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851352 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851434 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-config\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851462 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851482 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd5t5\" (UniqueName: \"kubernetes.io/projected/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-kube-api-access-bd5t5\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851519 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851548 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prtdj\" (UniqueName: \"kubernetes.io/projected/9c6f6e08-5872-47ea-86d3-a185f90d981d-kube-api-access-prtdj\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851569 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6f6e08-5872-47ea-86d3-a185f90d981d-logs\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851594 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851611 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851650 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-config-data\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.851668 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.852643 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6f6e08-5872-47ea-86d3-a185f90d981d-logs\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.863799 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.865554 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-config-data\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.869756 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prtdj\" (UniqueName: \"kubernetes.io/projected/9c6f6e08-5872-47ea-86d3-a185f90d981d-kube-api-access-prtdj\") pod \"nova-metadata-0\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " pod="openstack/nova-metadata-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.953918 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.953988 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954018 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b28nk\" (UniqueName: \"kubernetes.io/projected/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-kube-api-access-b28nk\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954075 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954174 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-config\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954231 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954256 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd5t5\" (UniqueName: \"kubernetes.io/projected/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-kube-api-access-bd5t5\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954322 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.954340 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.955411 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.955948 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.956578 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-config\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.957096 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.957648 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-svc\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.959051 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.962856 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.975539 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd5t5\" (UniqueName: \"kubernetes.io/projected/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-kube-api-access-bd5t5\") pod \"dnsmasq-dns-757b4f8459-gplfb\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:42 crc kubenswrapper[4829]: I1002 07:36:42.979174 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b28nk\" (UniqueName: \"kubernetes.io/projected/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-kube-api-access-b28nk\") pod \"nova-cell1-novncproxy-0\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.015710 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.028362 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.061316 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.071710 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.207754 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4jqqc"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.294854 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4jqqc" event={"ID":"0f195e99-4316-4030-9e49-1bf368defe41","Type":"ContainerStarted","Data":"a14abb21eb5ef36c3740c18b0226288bd9c223b1b66af5add34930cc1aff5f43"} Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.338311 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlwp8"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.339501 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.343971 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.344162 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.365446 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlwp8"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.429586 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.477363 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-scripts\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.477647 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.477825 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-config-data\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.477937 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8tzm\" (UniqueName: \"kubernetes.io/projected/d582f285-53ae-41fd-9be1-9b18dd9e95f9-kube-api-access-f8tzm\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.498785 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.546159 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.579255 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-scripts\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.579321 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.580126 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-config-data\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.580155 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8tzm\" (UniqueName: \"kubernetes.io/projected/d582f285-53ae-41fd-9be1-9b18dd9e95f9-kube-api-access-f8tzm\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.586676 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.587166 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-scripts\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.589836 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-config-data\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.598213 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8tzm\" (UniqueName: \"kubernetes.io/projected/d582f285-53ae-41fd-9be1-9b18dd9e95f9-kube-api-access-f8tzm\") pod \"nova-cell1-conductor-db-sync-zlwp8\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.714724 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.721304 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.816972 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:36:43 crc kubenswrapper[4829]: I1002 07:36:43.826094 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gplfb"] Oct 02 07:36:43 crc kubenswrapper[4829]: W1002 07:36:43.832338 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc720e5ad_6e31_43ce_b7c1_a0b590e58a90.slice/crio-92bcead0765f53d51cb35e17aa73955f356e91231b50b72b68b939e735b0f16b WatchSource:0}: Error finding container 92bcead0765f53d51cb35e17aa73955f356e91231b50b72b68b939e735b0f16b: Status 404 returned error can't find the container with id 92bcead0765f53d51cb35e17aa73955f356e91231b50b72b68b939e735b0f16b Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.245487 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlwp8"] Oct 02 07:36:44 crc kubenswrapper[4829]: W1002 07:36:44.265743 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd582f285_53ae_41fd_9be1_9b18dd9e95f9.slice/crio-1383dd908a198061eabdbfac6d903006a4c7c180e4b1884bffee4e9c5bc33c97 WatchSource:0}: Error finding container 1383dd908a198061eabdbfac6d903006a4c7c180e4b1884bffee4e9c5bc33c97: Status 404 returned error can't find the container with id 1383dd908a198061eabdbfac6d903006a4c7c180e4b1884bffee4e9c5bc33c97 Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.308007 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" event={"ID":"d582f285-53ae-41fd-9be1-9b18dd9e95f9","Type":"ContainerStarted","Data":"1383dd908a198061eabdbfac6d903006a4c7c180e4b1884bffee4e9c5bc33c97"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.309353 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9c6f6e08-5872-47ea-86d3-a185f90d981d","Type":"ContainerStarted","Data":"e27e741464a6364d6120bdd51e05b42cc25000cf07f581f96898a93132e3b22d"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.319799 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c720e5ad-6e31-43ce-b7c1-a0b590e58a90","Type":"ContainerStarted","Data":"92bcead0765f53d51cb35e17aa73955f356e91231b50b72b68b939e735b0f16b"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.323500 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096","Type":"ContainerStarted","Data":"1ac4e634d1f0b2b280f783adc2078759d9018f58e4c33a2968528868c76da78d"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.326010 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" event={"ID":"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121","Type":"ContainerStarted","Data":"d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.326035 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" event={"ID":"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121","Type":"ContainerStarted","Data":"fc2d33fd17cfc1fdf2890079101e39c32176fd134d215951541a8bcf7be5a4d7"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.336039 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"958f85b0-6cda-4f95-b713-4b0b59237d24","Type":"ContainerStarted","Data":"470bd2195e9be67094673bbcf7ef3bb7cf86dd4b50f2a8d077866168a517c157"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.340975 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4jqqc" event={"ID":"0f195e99-4316-4030-9e49-1bf368defe41","Type":"ContainerStarted","Data":"160849ffc45791c87a99b94a6fbd5c188e073902530fd11a5276ddd64b119939"} Oct 02 07:36:44 crc kubenswrapper[4829]: I1002 07:36:44.373163 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-4jqqc" podStartSLOduration=2.373145376 podStartE2EDuration="2.373145376s" podCreationTimestamp="2025-10-02 07:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:44.368921338 +0000 UTC m=+1195.708569743" watchObservedRunningTime="2025-10-02 07:36:44.373145376 +0000 UTC m=+1195.712793781" Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.358682 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" event={"ID":"d582f285-53ae-41fd-9be1-9b18dd9e95f9","Type":"ContainerStarted","Data":"ba9fb13ff51c68b72c7b7787afa0b55465ee90a6af58d91e5626e2287e81a1ed"} Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.361245 4829 generic.go:334] "Generic (PLEG): container finished" podID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerID="d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be" exitCode=0 Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.361313 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" event={"ID":"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121","Type":"ContainerDied","Data":"d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be"} Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.361342 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" event={"ID":"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121","Type":"ContainerStarted","Data":"c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970"} Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.361508 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.382880 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" podStartSLOduration=2.382861338 podStartE2EDuration="2.382861338s" podCreationTimestamp="2025-10-02 07:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:45.374766085 +0000 UTC m=+1196.714414500" watchObservedRunningTime="2025-10-02 07:36:45.382861338 +0000 UTC m=+1196.722509743" Oct 02 07:36:45 crc kubenswrapper[4829]: I1002 07:36:45.403893 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" podStartSLOduration=3.403868479 podStartE2EDuration="3.403868479s" podCreationTimestamp="2025-10-02 07:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:45.39461909 +0000 UTC m=+1196.734267495" watchObservedRunningTime="2025-10-02 07:36:45.403868479 +0000 UTC m=+1196.743516894" Oct 02 07:36:46 crc kubenswrapper[4829]: I1002 07:36:46.365335 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:36:46 crc kubenswrapper[4829]: I1002 07:36:46.381720 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.391255 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"958f85b0-6cda-4f95-b713-4b0b59237d24","Type":"ContainerStarted","Data":"aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2"} Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.393613 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9c6f6e08-5872-47ea-86d3-a185f90d981d","Type":"ContainerStarted","Data":"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f"} Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.393638 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9c6f6e08-5872-47ea-86d3-a185f90d981d","Type":"ContainerStarted","Data":"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc"} Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.393707 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-log" containerID="cri-o://cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc" gracePeriod=30 Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.393747 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-metadata" containerID="cri-o://ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f" gracePeriod=30 Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.395835 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c720e5ad-6e31-43ce-b7c1-a0b590e58a90","Type":"ContainerStarted","Data":"94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661"} Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.395965 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="c720e5ad-6e31-43ce-b7c1-a0b590e58a90" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661" gracePeriod=30 Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.404861 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096","Type":"ContainerStarted","Data":"a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866"} Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.405072 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096","Type":"ContainerStarted","Data":"6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254"} Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.428440 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.667181237 podStartE2EDuration="6.428421467s" podCreationTimestamp="2025-10-02 07:36:42 +0000 UTC" firstStartedPulling="2025-10-02 07:36:43.547276423 +0000 UTC m=+1194.886924828" lastFinishedPulling="2025-10-02 07:36:47.308516643 +0000 UTC m=+1198.648165058" observedRunningTime="2025-10-02 07:36:48.424023645 +0000 UTC m=+1199.763672050" watchObservedRunningTime="2025-10-02 07:36:48.428421467 +0000 UTC m=+1199.768069872" Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.472061 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9308210949999998 podStartE2EDuration="6.472042602s" podCreationTimestamp="2025-10-02 07:36:42 +0000 UTC" firstStartedPulling="2025-10-02 07:36:43.767375428 +0000 UTC m=+1195.107023833" lastFinishedPulling="2025-10-02 07:36:47.308596935 +0000 UTC m=+1198.648245340" observedRunningTime="2025-10-02 07:36:48.46766617 +0000 UTC m=+1199.807314565" watchObservedRunningTime="2025-10-02 07:36:48.472042602 +0000 UTC m=+1199.811691017" Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.472741 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.9999447740000003 podStartE2EDuration="6.472735544s" podCreationTimestamp="2025-10-02 07:36:42 +0000 UTC" firstStartedPulling="2025-10-02 07:36:43.837679197 +0000 UTC m=+1195.177327602" lastFinishedPulling="2025-10-02 07:36:47.310469967 +0000 UTC m=+1198.650118372" observedRunningTime="2025-10-02 07:36:48.447568268 +0000 UTC m=+1199.787216683" watchObservedRunningTime="2025-10-02 07:36:48.472735544 +0000 UTC m=+1199.812383949" Oct 02 07:36:48 crc kubenswrapper[4829]: I1002 07:36:48.497212 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.602902813 podStartE2EDuration="6.497190597s" podCreationTimestamp="2025-10-02 07:36:42 +0000 UTC" firstStartedPulling="2025-10-02 07:36:43.414636983 +0000 UTC m=+1194.754285388" lastFinishedPulling="2025-10-02 07:36:47.308924767 +0000 UTC m=+1198.648573172" observedRunningTime="2025-10-02 07:36:48.48833307 +0000 UTC m=+1199.827981465" watchObservedRunningTime="2025-10-02 07:36:48.497190597 +0000 UTC m=+1199.836839002" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.019240 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.021947 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prtdj\" (UniqueName: \"kubernetes.io/projected/9c6f6e08-5872-47ea-86d3-a185f90d981d-kube-api-access-prtdj\") pod \"9c6f6e08-5872-47ea-86d3-a185f90d981d\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.022145 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-combined-ca-bundle\") pod \"9c6f6e08-5872-47ea-86d3-a185f90d981d\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.022209 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-config-data\") pod \"9c6f6e08-5872-47ea-86d3-a185f90d981d\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.022289 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6f6e08-5872-47ea-86d3-a185f90d981d-logs\") pod \"9c6f6e08-5872-47ea-86d3-a185f90d981d\" (UID: \"9c6f6e08-5872-47ea-86d3-a185f90d981d\") " Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.022690 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c6f6e08-5872-47ea-86d3-a185f90d981d-logs" (OuterVolumeSpecName: "logs") pod "9c6f6e08-5872-47ea-86d3-a185f90d981d" (UID: "9c6f6e08-5872-47ea-86d3-a185f90d981d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.023026 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6f6e08-5872-47ea-86d3-a185f90d981d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.027525 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c6f6e08-5872-47ea-86d3-a185f90d981d-kube-api-access-prtdj" (OuterVolumeSpecName: "kube-api-access-prtdj") pod "9c6f6e08-5872-47ea-86d3-a185f90d981d" (UID: "9c6f6e08-5872-47ea-86d3-a185f90d981d"). InnerVolumeSpecName "kube-api-access-prtdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.066450 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-config-data" (OuterVolumeSpecName: "config-data") pod "9c6f6e08-5872-47ea-86d3-a185f90d981d" (UID: "9c6f6e08-5872-47ea-86d3-a185f90d981d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.087242 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c6f6e08-5872-47ea-86d3-a185f90d981d" (UID: "9c6f6e08-5872-47ea-86d3-a185f90d981d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.123888 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prtdj\" (UniqueName: \"kubernetes.io/projected/9c6f6e08-5872-47ea-86d3-a185f90d981d-kube-api-access-prtdj\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.123919 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.123928 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6f6e08-5872-47ea-86d3-a185f90d981d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.382700 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.422938 4829 generic.go:334] "Generic (PLEG): container finished" podID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerID="ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f" exitCode=0 Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.422976 4829 generic.go:334] "Generic (PLEG): container finished" podID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerID="cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc" exitCode=143 Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.422985 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.423000 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9c6f6e08-5872-47ea-86d3-a185f90d981d","Type":"ContainerDied","Data":"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f"} Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.423042 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9c6f6e08-5872-47ea-86d3-a185f90d981d","Type":"ContainerDied","Data":"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc"} Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.432483 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9c6f6e08-5872-47ea-86d3-a185f90d981d","Type":"ContainerDied","Data":"e27e741464a6364d6120bdd51e05b42cc25000cf07f581f96898a93132e3b22d"} Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.423100 4829 scope.go:117] "RemoveContainer" containerID="ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.469739 4829 scope.go:117] "RemoveContainer" containerID="cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.511106 4829 scope.go:117] "RemoveContainer" containerID="ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f" Oct 02 07:36:49 crc kubenswrapper[4829]: E1002 07:36:49.526627 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f\": container with ID starting with ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f not found: ID does not exist" containerID="ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.526689 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f"} err="failed to get container status \"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f\": rpc error: code = NotFound desc = could not find container \"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f\": container with ID starting with ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f not found: ID does not exist" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.526720 4829 scope.go:117] "RemoveContainer" containerID="cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc" Oct 02 07:36:49 crc kubenswrapper[4829]: E1002 07:36:49.527663 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc\": container with ID starting with cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc not found: ID does not exist" containerID="cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.527682 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc"} err="failed to get container status \"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc\": rpc error: code = NotFound desc = could not find container \"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc\": container with ID starting with cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc not found: ID does not exist" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.527701 4829 scope.go:117] "RemoveContainer" containerID="ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.528644 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f"} err="failed to get container status \"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f\": rpc error: code = NotFound desc = could not find container \"ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f\": container with ID starting with ca8b947d36c787b4182172e6f04c29ea16b0df6a62c7155f27373bd243564e9f not found: ID does not exist" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.528667 4829 scope.go:117] "RemoveContainer" containerID="cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.529066 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc"} err="failed to get container status \"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc\": rpc error: code = NotFound desc = could not find container \"cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc\": container with ID starting with cdf95617eebf9a94979b15a4592993c5b8c80528c5015e6eb8a5f5c9406562bc not found: ID does not exist" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.590511 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.617038 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.629476 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:49 crc kubenswrapper[4829]: E1002 07:36:49.630657 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-log" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.630683 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-log" Oct 02 07:36:49 crc kubenswrapper[4829]: E1002 07:36:49.630697 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-metadata" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.630703 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-metadata" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.631060 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-metadata" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.631076 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" containerName="nova-metadata-log" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.632130 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.638587 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.638764 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.642386 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.651797 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.651876 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.651959 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-config-data\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.652050 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19bfc874-e2e6-462e-a0da-15fe04932f9e-logs\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.652104 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kz5v\" (UniqueName: \"kubernetes.io/projected/19bfc874-e2e6-462e-a0da-15fe04932f9e-kube-api-access-6kz5v\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.753344 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19bfc874-e2e6-462e-a0da-15fe04932f9e-logs\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.753710 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kz5v\" (UniqueName: \"kubernetes.io/projected/19bfc874-e2e6-462e-a0da-15fe04932f9e-kube-api-access-6kz5v\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.753837 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.753946 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.753856 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19bfc874-e2e6-462e-a0da-15fe04932f9e-logs\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.754814 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-config-data\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.757607 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.758130 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.758889 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-config-data\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.778038 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kz5v\" (UniqueName: \"kubernetes.io/projected/19bfc874-e2e6-462e-a0da-15fe04932f9e-kube-api-access-6kz5v\") pod \"nova-metadata-0\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " pod="openstack/nova-metadata-0" Oct 02 07:36:49 crc kubenswrapper[4829]: I1002 07:36:49.955878 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:50 crc kubenswrapper[4829]: I1002 07:36:50.527038 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.475142 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c6f6e08-5872-47ea-86d3-a185f90d981d" path="/var/lib/kubelet/pods/9c6f6e08-5872-47ea-86d3-a185f90d981d/volumes" Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.501544 4829 generic.go:334] "Generic (PLEG): container finished" podID="0f195e99-4316-4030-9e49-1bf368defe41" containerID="160849ffc45791c87a99b94a6fbd5c188e073902530fd11a5276ddd64b119939" exitCode=0 Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.501612 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4jqqc" event={"ID":"0f195e99-4316-4030-9e49-1bf368defe41","Type":"ContainerDied","Data":"160849ffc45791c87a99b94a6fbd5c188e073902530fd11a5276ddd64b119939"} Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.505050 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19bfc874-e2e6-462e-a0da-15fe04932f9e","Type":"ContainerStarted","Data":"f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385"} Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.505086 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19bfc874-e2e6-462e-a0da-15fe04932f9e","Type":"ContainerStarted","Data":"956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d"} Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.505100 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19bfc874-e2e6-462e-a0da-15fe04932f9e","Type":"ContainerStarted","Data":"1966a00ee1d9cb2ed6e2c4d4f7bd41fcdcd2b359e83d49fd82c0fd0f31761cc2"} Oct 02 07:36:51 crc kubenswrapper[4829]: I1002 07:36:51.553590 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.553568177 podStartE2EDuration="2.553568177s" podCreationTimestamp="2025-10-02 07:36:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:51.549890968 +0000 UTC m=+1202.889539373" watchObservedRunningTime="2025-10-02 07:36:51.553568177 +0000 UTC m=+1202.893216572" Oct 02 07:36:52 crc kubenswrapper[4829]: I1002 07:36:52.526795 4829 generic.go:334] "Generic (PLEG): container finished" podID="d582f285-53ae-41fd-9be1-9b18dd9e95f9" containerID="ba9fb13ff51c68b72c7b7787afa0b55465ee90a6af58d91e5626e2287e81a1ed" exitCode=0 Oct 02 07:36:52 crc kubenswrapper[4829]: I1002 07:36:52.526868 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" event={"ID":"d582f285-53ae-41fd-9be1-9b18dd9e95f9","Type":"ContainerDied","Data":"ba9fb13ff51c68b72c7b7787afa0b55465ee90a6af58d91e5626e2287e81a1ed"} Oct 02 07:36:52 crc kubenswrapper[4829]: I1002 07:36:52.720872 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:36:52 crc kubenswrapper[4829]: I1002 07:36:52.720918 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:36:52 crc kubenswrapper[4829]: I1002 07:36:52.999114 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.016371 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.017539 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.029898 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr9xl\" (UniqueName: \"kubernetes.io/projected/0f195e99-4316-4030-9e49-1bf368defe41-kube-api-access-vr9xl\") pod \"0f195e99-4316-4030-9e49-1bf368defe41\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.029943 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-config-data\") pod \"0f195e99-4316-4030-9e49-1bf368defe41\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.030106 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-combined-ca-bundle\") pod \"0f195e99-4316-4030-9e49-1bf368defe41\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.030213 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-scripts\") pod \"0f195e99-4316-4030-9e49-1bf368defe41\" (UID: \"0f195e99-4316-4030-9e49-1bf368defe41\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.043485 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-scripts" (OuterVolumeSpecName: "scripts") pod "0f195e99-4316-4030-9e49-1bf368defe41" (UID: "0f195e99-4316-4030-9e49-1bf368defe41"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.044390 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f195e99-4316-4030-9e49-1bf368defe41-kube-api-access-vr9xl" (OuterVolumeSpecName: "kube-api-access-vr9xl") pod "0f195e99-4316-4030-9e49-1bf368defe41" (UID: "0f195e99-4316-4030-9e49-1bf368defe41"). InnerVolumeSpecName "kube-api-access-vr9xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.061474 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.062867 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.071766 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f195e99-4316-4030-9e49-1bf368defe41" (UID: "0f195e99-4316-4030-9e49-1bf368defe41"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.085896 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.120430 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-config-data" (OuterVolumeSpecName: "config-data") pod "0f195e99-4316-4030-9e49-1bf368defe41" (UID: "0f195e99-4316-4030-9e49-1bf368defe41"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.132403 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.132430 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.132440 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f195e99-4316-4030-9e49-1bf368defe41-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.132447 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr9xl\" (UniqueName: \"kubernetes.io/projected/0f195e99-4316-4030-9e49-1bf368defe41-kube-api-access-vr9xl\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.159781 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cmmvh"] Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.159999 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="dnsmasq-dns" containerID="cri-o://3c52ebe8e796e8bca72e4cb68184e3949bbee197dd0435c2787e4c7aa2243eb3" gracePeriod=10 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.346879 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.178:5353: connect: connection refused" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.543639 4829 generic.go:334] "Generic (PLEG): container finished" podID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerID="3c52ebe8e796e8bca72e4cb68184e3949bbee197dd0435c2787e4c7aa2243eb3" exitCode=0 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.543699 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" event={"ID":"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8","Type":"ContainerDied","Data":"3c52ebe8e796e8bca72e4cb68184e3949bbee197dd0435c2787e4c7aa2243eb3"} Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.555067 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4jqqc" event={"ID":"0f195e99-4316-4030-9e49-1bf368defe41","Type":"ContainerDied","Data":"a14abb21eb5ef36c3740c18b0226288bd9c223b1b66af5add34930cc1aff5f43"} Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.555418 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a14abb21eb5ef36c3740c18b0226288bd9c223b1b66af5add34930cc1aff5f43" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.555217 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4jqqc" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.589439 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.686874 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.744802 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-sb\") pod \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.744993 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-swift-storage-0\") pod \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.745080 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-svc\") pod \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.745110 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-nb\") pod \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.745127 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-config\") pod \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.745212 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x9hh\" (UniqueName: \"kubernetes.io/projected/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-kube-api-access-4x9hh\") pod \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\" (UID: \"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8\") " Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.750636 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-kube-api-access-4x9hh" (OuterVolumeSpecName: "kube-api-access-4x9hh") pod "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" (UID: "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8"). InnerVolumeSpecName "kube-api-access-4x9hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.762419 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.762692 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.202:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.768294 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.768526 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-log" containerID="cri-o://6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254" gracePeriod=30 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.768905 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-api" containerID="cri-o://a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866" gracePeriod=30 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.782852 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.783057 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-log" containerID="cri-o://956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d" gracePeriod=30 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.783460 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-metadata" containerID="cri-o://f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385" gracePeriod=30 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.837271 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" (UID: "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.841976 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" (UID: "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.857456 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.857492 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x9hh\" (UniqueName: \"kubernetes.io/projected/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-kube-api-access-4x9hh\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.857504 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.865720 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" (UID: "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.900759 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" (UID: "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.947832 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.948038 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" containerName="kube-state-metrics" containerID="cri-o://76959c27443e03cc3089c6da40ef4e114f4c5d9bd80af46e12cb7814d888978f" gracePeriod=30 Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.952894 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-config" (OuterVolumeSpecName: "config") pod "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" (UID: "c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.959736 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.959769 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:53 crc kubenswrapper[4829]: I1002 07:36:53.959777 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.175639 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.273030 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-config-data\") pod \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.273091 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-combined-ca-bundle\") pod \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.273125 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-scripts\") pod \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.273211 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8tzm\" (UniqueName: \"kubernetes.io/projected/d582f285-53ae-41fd-9be1-9b18dd9e95f9-kube-api-access-f8tzm\") pod \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\" (UID: \"d582f285-53ae-41fd-9be1-9b18dd9e95f9\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.278504 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-scripts" (OuterVolumeSpecName: "scripts") pod "d582f285-53ae-41fd-9be1-9b18dd9e95f9" (UID: "d582f285-53ae-41fd-9be1-9b18dd9e95f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.278820 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d582f285-53ae-41fd-9be1-9b18dd9e95f9-kube-api-access-f8tzm" (OuterVolumeSpecName: "kube-api-access-f8tzm") pod "d582f285-53ae-41fd-9be1-9b18dd9e95f9" (UID: "d582f285-53ae-41fd-9be1-9b18dd9e95f9"). InnerVolumeSpecName "kube-api-access-f8tzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.304826 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d582f285-53ae-41fd-9be1-9b18dd9e95f9" (UID: "d582f285-53ae-41fd-9be1-9b18dd9e95f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.311311 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-config-data" (OuterVolumeSpecName: "config-data") pod "d582f285-53ae-41fd-9be1-9b18dd9e95f9" (UID: "d582f285-53ae-41fd-9be1-9b18dd9e95f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.375546 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8tzm\" (UniqueName: \"kubernetes.io/projected/d582f285-53ae-41fd-9be1-9b18dd9e95f9-kube-api-access-f8tzm\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.375582 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.375591 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.375600 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d582f285-53ae-41fd-9be1-9b18dd9e95f9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.555301 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.565113 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" event={"ID":"d582f285-53ae-41fd-9be1-9b18dd9e95f9","Type":"ContainerDied","Data":"1383dd908a198061eabdbfac6d903006a4c7c180e4b1884bffee4e9c5bc33c97"} Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.565173 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1383dd908a198061eabdbfac6d903006a4c7c180e4b1884bffee4e9c5bc33c97" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.565261 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-zlwp8" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.583536 4829 generic.go:334] "Generic (PLEG): container finished" podID="088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" containerID="76959c27443e03cc3089c6da40ef4e114f4c5d9bd80af46e12cb7814d888978f" exitCode=2 Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.583826 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3","Type":"ContainerDied","Data":"76959c27443e03cc3089c6da40ef4e114f4c5d9bd80af46e12cb7814d888978f"} Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.592979 4829 generic.go:334] "Generic (PLEG): container finished" podID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerID="f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385" exitCode=0 Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.593183 4829 generic.go:334] "Generic (PLEG): container finished" podID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerID="956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d" exitCode=143 Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.593329 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19bfc874-e2e6-462e-a0da-15fe04932f9e","Type":"ContainerDied","Data":"f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385"} Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.593419 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19bfc874-e2e6-462e-a0da-15fe04932f9e","Type":"ContainerDied","Data":"956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d"} Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.594338 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.608052 4829 generic.go:334] "Generic (PLEG): container finished" podID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerID="6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254" exitCode=143 Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.608311 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096","Type":"ContainerDied","Data":"6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254"} Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.623342 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.623959 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cmmvh" event={"ID":"c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8","Type":"ContainerDied","Data":"c0dceb5d376687ff56e5fd5a20170b4e60bd7c4971b247c51f8bed97ed1ff692"} Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.624118 4829 scope.go:117] "RemoveContainer" containerID="3c52ebe8e796e8bca72e4cb68184e3949bbee197dd0435c2787e4c7aa2243eb3" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.645648 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 07:36:54 crc kubenswrapper[4829]: E1002 07:36:54.646257 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f195e99-4316-4030-9e49-1bf368defe41" containerName="nova-manage" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646269 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f195e99-4316-4030-9e49-1bf368defe41" containerName="nova-manage" Oct 02 07:36:54 crc kubenswrapper[4829]: E1002 07:36:54.646281 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d582f285-53ae-41fd-9be1-9b18dd9e95f9" containerName="nova-cell1-conductor-db-sync" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646287 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d582f285-53ae-41fd-9be1-9b18dd9e95f9" containerName="nova-cell1-conductor-db-sync" Oct 02 07:36:54 crc kubenswrapper[4829]: E1002 07:36:54.646305 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="init" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646311 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="init" Oct 02 07:36:54 crc kubenswrapper[4829]: E1002 07:36:54.646328 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-log" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646333 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-log" Oct 02 07:36:54 crc kubenswrapper[4829]: E1002 07:36:54.646349 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-metadata" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646355 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-metadata" Oct 02 07:36:54 crc kubenswrapper[4829]: E1002 07:36:54.646379 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="dnsmasq-dns" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646385 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="dnsmasq-dns" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646545 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d582f285-53ae-41fd-9be1-9b18dd9e95f9" containerName="nova-cell1-conductor-db-sync" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646555 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" containerName="dnsmasq-dns" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646570 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-log" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646584 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f195e99-4316-4030-9e49-1bf368defe41" containerName="nova-manage" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.646598 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" containerName="nova-metadata-metadata" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.647411 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.651983 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.659007 4829 scope.go:117] "RemoveContainer" containerID="2040e619f09be56dae02ac8b949b77cfe20706963aac53700378978acbfe1f27" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.660283 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.684889 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-combined-ca-bundle\") pod \"19bfc874-e2e6-462e-a0da-15fe04932f9e\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.684971 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kz5v\" (UniqueName: \"kubernetes.io/projected/19bfc874-e2e6-462e-a0da-15fe04932f9e-kube-api-access-6kz5v\") pod \"19bfc874-e2e6-462e-a0da-15fe04932f9e\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.685053 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19bfc874-e2e6-462e-a0da-15fe04932f9e-logs\") pod \"19bfc874-e2e6-462e-a0da-15fe04932f9e\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.685097 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-nova-metadata-tls-certs\") pod \"19bfc874-e2e6-462e-a0da-15fe04932f9e\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.685117 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-config-data\") pod \"19bfc874-e2e6-462e-a0da-15fe04932f9e\" (UID: \"19bfc874-e2e6-462e-a0da-15fe04932f9e\") " Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.685450 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.685509 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7glc\" (UniqueName: \"kubernetes.io/projected/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-kube-api-access-q7glc\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.685555 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.691686 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19bfc874-e2e6-462e-a0da-15fe04932f9e-kube-api-access-6kz5v" (OuterVolumeSpecName: "kube-api-access-6kz5v") pod "19bfc874-e2e6-462e-a0da-15fe04932f9e" (UID: "19bfc874-e2e6-462e-a0da-15fe04932f9e"). InnerVolumeSpecName "kube-api-access-6kz5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.691719 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19bfc874-e2e6-462e-a0da-15fe04932f9e-logs" (OuterVolumeSpecName: "logs") pod "19bfc874-e2e6-462e-a0da-15fe04932f9e" (UID: "19bfc874-e2e6-462e-a0da-15fe04932f9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.732237 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cmmvh"] Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.736098 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-config-data" (OuterVolumeSpecName: "config-data") pod "19bfc874-e2e6-462e-a0da-15fe04932f9e" (UID: "19bfc874-e2e6-462e-a0da-15fe04932f9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.759087 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cmmvh"] Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.776429 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19bfc874-e2e6-462e-a0da-15fe04932f9e" (UID: "19bfc874-e2e6-462e-a0da-15fe04932f9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.785375 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "19bfc874-e2e6-462e-a0da-15fe04932f9e" (UID: "19bfc874-e2e6-462e-a0da-15fe04932f9e"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.786945 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787023 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7glc\" (UniqueName: \"kubernetes.io/projected/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-kube-api-access-q7glc\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787061 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787149 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787162 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kz5v\" (UniqueName: \"kubernetes.io/projected/19bfc874-e2e6-462e-a0da-15fe04932f9e-kube-api-access-6kz5v\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787172 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19bfc874-e2e6-462e-a0da-15fe04932f9e-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787180 4829 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.787188 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19bfc874-e2e6-462e-a0da-15fe04932f9e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.789920 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.797011 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.805057 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7glc\" (UniqueName: \"kubernetes.io/projected/483a24aa-87d3-4b0a-accd-5ca4a8a7a029-kube-api-access-q7glc\") pod \"nova-cell1-conductor-0\" (UID: \"483a24aa-87d3-4b0a-accd-5ca4a8a7a029\") " pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:54 crc kubenswrapper[4829]: I1002 07:36:54.979232 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.164106 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.192594 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs9h5\" (UniqueName: \"kubernetes.io/projected/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3-kube-api-access-xs9h5\") pod \"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3\" (UID: \"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3\") " Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.200350 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3-kube-api-access-xs9h5" (OuterVolumeSpecName: "kube-api-access-xs9h5") pod "088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" (UID: "088860ca-f3ac-4f1b-86ae-13a3a13a8ba3"). InnerVolumeSpecName "kube-api-access-xs9h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.294632 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs9h5\" (UniqueName: \"kubernetes.io/projected/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3-kube-api-access-xs9h5\") on node \"crc\" DevicePath \"\"" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.485675 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8" path="/var/lib/kubelet/pods/c7d5d635-21f5-42d1-8ea8-3b7f9be2c0e8/volumes" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.548985 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.640498 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"088860ca-f3ac-4f1b-86ae-13a3a13a8ba3","Type":"ContainerDied","Data":"5c97ef80661509fa86c0121858c0373f14492b2c5af443228ac222800a22d06e"} Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.640549 4829 scope.go:117] "RemoveContainer" containerID="76959c27443e03cc3089c6da40ef4e114f4c5d9bd80af46e12cb7814d888978f" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.640642 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.645770 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.646689 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"19bfc874-e2e6-462e-a0da-15fe04932f9e","Type":"ContainerDied","Data":"1966a00ee1d9cb2ed6e2c4d4f7bd41fcdcd2b359e83d49fd82c0fd0f31761cc2"} Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.648583 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="958f85b0-6cda-4f95-b713-4b0b59237d24" containerName="nova-scheduler-scheduler" containerID="cri-o://aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2" gracePeriod=30 Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.648961 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"483a24aa-87d3-4b0a-accd-5ca4a8a7a029","Type":"ContainerStarted","Data":"b88265e34a5fd71a35d3d1a7fb9ffb9a150e59ddf12c82b4e24ea5102d97585b"} Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.684258 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.697951 4829 scope.go:117] "RemoveContainer" containerID="f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.708598 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.737300 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: E1002 07:36:55.737771 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" containerName="kube-state-metrics" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.737784 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" containerName="kube-state-metrics" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.737967 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" containerName="kube-state-metrics" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.738835 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.744168 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.744376 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.749794 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.756106 4829 scope.go:117] "RemoveContainer" containerID="956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.760557 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.770096 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.778510 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.780280 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.782086 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.782543 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.788794 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.827430 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-config-data\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.827506 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.827536 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m75xp\" (UniqueName: \"kubernetes.io/projected/fadeaa43-1547-4915-9e34-43e880485d84-kube-api-access-m75xp\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.827571 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66ddh\" (UniqueName: \"kubernetes.io/projected/88ecb52c-ee97-45fc-9701-69d9d2abf67f-kube-api-access-66ddh\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.827729 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.827893 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.828109 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.828186 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ecb52c-ee97-45fc-9701-69d9d2abf67f-logs\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.828399 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929738 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m75xp\" (UniqueName: \"kubernetes.io/projected/fadeaa43-1547-4915-9e34-43e880485d84-kube-api-access-m75xp\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929795 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66ddh\" (UniqueName: \"kubernetes.io/projected/88ecb52c-ee97-45fc-9701-69d9d2abf67f-kube-api-access-66ddh\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929823 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929869 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929931 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929948 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ecb52c-ee97-45fc-9701-69d9d2abf67f-logs\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.929988 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.930021 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-config-data\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.930052 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.930885 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ecb52c-ee97-45fc-9701-69d9d2abf67f-logs\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.935013 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.935345 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.935510 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-config-data\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.941014 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.941615 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.944422 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fadeaa43-1547-4915-9e34-43e880485d84-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.950866 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m75xp\" (UniqueName: \"kubernetes.io/projected/fadeaa43-1547-4915-9e34-43e880485d84-kube-api-access-m75xp\") pod \"kube-state-metrics-0\" (UID: \"fadeaa43-1547-4915-9e34-43e880485d84\") " pod="openstack/kube-state-metrics-0" Oct 02 07:36:55 crc kubenswrapper[4829]: I1002 07:36:55.954021 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66ddh\" (UniqueName: \"kubernetes.io/projected/88ecb52c-ee97-45fc-9701-69d9d2abf67f-kube-api-access-66ddh\") pod \"nova-metadata-0\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " pod="openstack/nova-metadata-0" Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.066242 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.098755 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.624684 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.669208 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fadeaa43-1547-4915-9e34-43e880485d84","Type":"ContainerStarted","Data":"12d67fa3b4a6e43c5cd186484d8b1917d9378398dc883420c77e4aa108a375e1"} Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.671675 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"483a24aa-87d3-4b0a-accd-5ca4a8a7a029","Type":"ContainerStarted","Data":"c89c974d1ed92d4ba7fcbf58a15ac441b9beea7f2a43dd1b8d55a205b888e429"} Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.671830 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.691018 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.690994239 podStartE2EDuration="2.690994239s" podCreationTimestamp="2025-10-02 07:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:56.686978179 +0000 UTC m=+1208.026626594" watchObservedRunningTime="2025-10-02 07:36:56.690994239 +0000 UTC m=+1208.030642644" Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.813563 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.889135 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.889654 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-central-agent" containerID="cri-o://13f6ccd385bcbd94cc65b2964ca612bf866091f1e4f8a9526fe7b83dec51ceb4" gracePeriod=30 Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.889701 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="proxy-httpd" containerID="cri-o://541258c680fb40bdf3dfe44364e81d92bf3100c73c2bdbf3cc519b48c3673666" gracePeriod=30 Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.889820 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-notification-agent" containerID="cri-o://b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72" gracePeriod=30 Oct 02 07:36:56 crc kubenswrapper[4829]: I1002 07:36:56.889825 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="sg-core" containerID="cri-o://4fbb90c51caee159cef1056dd8484fa71b4577db9ac880c9dc1c443aa3abb4f5" gracePeriod=30 Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.473357 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="088860ca-f3ac-4f1b-86ae-13a3a13a8ba3" path="/var/lib/kubelet/pods/088860ca-f3ac-4f1b-86ae-13a3a13a8ba3/volumes" Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.474591 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19bfc874-e2e6-462e-a0da-15fe04932f9e" path="/var/lib/kubelet/pods/19bfc874-e2e6-462e-a0da-15fe04932f9e/volumes" Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.683795 4829 generic.go:334] "Generic (PLEG): container finished" podID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerID="541258c680fb40bdf3dfe44364e81d92bf3100c73c2bdbf3cc519b48c3673666" exitCode=0 Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.683832 4829 generic.go:334] "Generic (PLEG): container finished" podID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerID="4fbb90c51caee159cef1056dd8484fa71b4577db9ac880c9dc1c443aa3abb4f5" exitCode=2 Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.683847 4829 generic.go:334] "Generic (PLEG): container finished" podID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerID="13f6ccd385bcbd94cc65b2964ca612bf866091f1e4f8a9526fe7b83dec51ceb4" exitCode=0 Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.683893 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerDied","Data":"541258c680fb40bdf3dfe44364e81d92bf3100c73c2bdbf3cc519b48c3673666"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.683923 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerDied","Data":"4fbb90c51caee159cef1056dd8484fa71b4577db9ac880c9dc1c443aa3abb4f5"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.683937 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerDied","Data":"13f6ccd385bcbd94cc65b2964ca612bf866091f1e4f8a9526fe7b83dec51ceb4"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.685715 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"fadeaa43-1547-4915-9e34-43e880485d84","Type":"ContainerStarted","Data":"773c217ceee97596a7662dc9d9b67c4fde6ab36d8cbe3cfae75fbfa39512b267"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.687441 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.689528 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88ecb52c-ee97-45fc-9701-69d9d2abf67f","Type":"ContainerStarted","Data":"01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.689568 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88ecb52c-ee97-45fc-9701-69d9d2abf67f","Type":"ContainerStarted","Data":"1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.689586 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88ecb52c-ee97-45fc-9701-69d9d2abf67f","Type":"ContainerStarted","Data":"c1fd68a6acbd5b72b5b50a1b68677203e1519b2de488bc9297855795cb17ac72"} Oct 02 07:36:57 crc kubenswrapper[4829]: I1002 07:36:57.713304 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.295195957 podStartE2EDuration="2.7132793s" podCreationTimestamp="2025-10-02 07:36:55 +0000 UTC" firstStartedPulling="2025-10-02 07:36:56.63702149 +0000 UTC m=+1207.976669895" lastFinishedPulling="2025-10-02 07:36:57.055104833 +0000 UTC m=+1208.394753238" observedRunningTime="2025-10-02 07:36:57.704695112 +0000 UTC m=+1209.044343527" watchObservedRunningTime="2025-10-02 07:36:57.7132793 +0000 UTC m=+1209.052927745" Oct 02 07:36:58 crc kubenswrapper[4829]: E1002 07:36:58.020390 4829 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 07:36:58 crc kubenswrapper[4829]: E1002 07:36:58.023889 4829 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 07:36:58 crc kubenswrapper[4829]: E1002 07:36:58.026026 4829 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 07:36:58 crc kubenswrapper[4829]: E1002 07:36:58.026063 4829 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="958f85b0-6cda-4f95-b713-4b0b59237d24" containerName="nova-scheduler-scheduler" Oct 02 07:36:59 crc kubenswrapper[4829]: I1002 07:36:59.740531 4829 generic.go:334] "Generic (PLEG): container finished" podID="958f85b0-6cda-4f95-b713-4b0b59237d24" containerID="aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2" exitCode=0 Oct 02 07:36:59 crc kubenswrapper[4829]: I1002 07:36:59.741357 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"958f85b0-6cda-4f95-b713-4b0b59237d24","Type":"ContainerDied","Data":"aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2"} Oct 02 07:36:59 crc kubenswrapper[4829]: I1002 07:36:59.928770 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:36:59 crc kubenswrapper[4829]: I1002 07:36:59.950237 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.950199715 podStartE2EDuration="4.950199715s" podCreationTimestamp="2025-10-02 07:36:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:36:57.729399983 +0000 UTC m=+1209.069048458" watchObservedRunningTime="2025-10-02 07:36:59.950199715 +0000 UTC m=+1211.289848130" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.018499 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-config-data\") pod \"958f85b0-6cda-4f95-b713-4b0b59237d24\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.019427 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-combined-ca-bundle\") pod \"958f85b0-6cda-4f95-b713-4b0b59237d24\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.019573 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnfth\" (UniqueName: \"kubernetes.io/projected/958f85b0-6cda-4f95-b713-4b0b59237d24-kube-api-access-qnfth\") pod \"958f85b0-6cda-4f95-b713-4b0b59237d24\" (UID: \"958f85b0-6cda-4f95-b713-4b0b59237d24\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.024387 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/958f85b0-6cda-4f95-b713-4b0b59237d24-kube-api-access-qnfth" (OuterVolumeSpecName: "kube-api-access-qnfth") pod "958f85b0-6cda-4f95-b713-4b0b59237d24" (UID: "958f85b0-6cda-4f95-b713-4b0b59237d24"). InnerVolumeSpecName "kube-api-access-qnfth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.058051 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-config-data" (OuterVolumeSpecName: "config-data") pod "958f85b0-6cda-4f95-b713-4b0b59237d24" (UID: "958f85b0-6cda-4f95-b713-4b0b59237d24"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.067033 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "958f85b0-6cda-4f95-b713-4b0b59237d24" (UID: "958f85b0-6cda-4f95-b713-4b0b59237d24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.123126 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.123482 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/958f85b0-6cda-4f95-b713-4b0b59237d24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.123507 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnfth\" (UniqueName: \"kubernetes.io/projected/958f85b0-6cda-4f95-b713-4b0b59237d24-kube-api-access-qnfth\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.700030 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.754340 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"958f85b0-6cda-4f95-b713-4b0b59237d24","Type":"ContainerDied","Data":"470bd2195e9be67094673bbcf7ef3bb7cf86dd4b50f2a8d077866168a517c157"} Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.754660 4829 scope.go:117] "RemoveContainer" containerID="aeb04c5f7e84be8f90244a9004bd9ddb19d0123def8ffa91d37104dce46ee3d2" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.754684 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.759490 4829 generic.go:334] "Generic (PLEG): container finished" podID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerID="a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866" exitCode=0 Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.759537 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096","Type":"ContainerDied","Data":"a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866"} Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.759564 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096","Type":"ContainerDied","Data":"1ac4e634d1f0b2b280f783adc2078759d9018f58e4c33a2968528868c76da78d"} Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.759667 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.771800 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-logs\") pod \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.771945 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-config-data\") pod \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.772018 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prtv9\" (UniqueName: \"kubernetes.io/projected/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-kube-api-access-prtv9\") pod \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.772084 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-combined-ca-bundle\") pod \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\" (UID: \"8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096\") " Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.773114 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-logs" (OuterVolumeSpecName: "logs") pod "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" (UID: "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.773926 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.788294 4829 scope.go:117] "RemoveContainer" containerID="a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.796708 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.804409 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.815283 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-kube-api-access-prtv9" (OuterVolumeSpecName: "kube-api-access-prtv9") pod "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" (UID: "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096"). InnerVolumeSpecName "kube-api-access-prtv9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.819414 4829 scope.go:117] "RemoveContainer" containerID="6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.825477 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-config-data" (OuterVolumeSpecName: "config-data") pod "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" (UID: "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.829937 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:00 crc kubenswrapper[4829]: E1002 07:37:00.830501 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="958f85b0-6cda-4f95-b713-4b0b59237d24" containerName="nova-scheduler-scheduler" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.830523 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="958f85b0-6cda-4f95-b713-4b0b59237d24" containerName="nova-scheduler-scheduler" Oct 02 07:37:00 crc kubenswrapper[4829]: E1002 07:37:00.830536 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-api" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.830543 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-api" Oct 02 07:37:00 crc kubenswrapper[4829]: E1002 07:37:00.830562 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-log" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.830569 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-log" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.830760 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="958f85b0-6cda-4f95-b713-4b0b59237d24" containerName="nova-scheduler-scheduler" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.830779 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-log" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.830793 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" containerName="nova-api-api" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.831618 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.831610 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" (UID: "8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.833703 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.860924 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.875117 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clvqb\" (UniqueName: \"kubernetes.io/projected/e9b54cbb-437d-4016-9ef7-8966ce4b9035-kube-api-access-clvqb\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.875154 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.875490 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-config-data\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.875812 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.875858 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prtv9\" (UniqueName: \"kubernetes.io/projected/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-kube-api-access-prtv9\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.875880 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.916217 4829 scope.go:117] "RemoveContainer" containerID="a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866" Oct 02 07:37:00 crc kubenswrapper[4829]: E1002 07:37:00.916747 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866\": container with ID starting with a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866 not found: ID does not exist" containerID="a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.916782 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866"} err="failed to get container status \"a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866\": rpc error: code = NotFound desc = could not find container \"a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866\": container with ID starting with a251a3481644b8fe2be099bab922dc578e231a6497e82859a7f6168571721866 not found: ID does not exist" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.916804 4829 scope.go:117] "RemoveContainer" containerID="6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254" Oct 02 07:37:00 crc kubenswrapper[4829]: E1002 07:37:00.917248 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254\": container with ID starting with 6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254 not found: ID does not exist" containerID="6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.917281 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254"} err="failed to get container status \"6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254\": rpc error: code = NotFound desc = could not find container \"6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254\": container with ID starting with 6bb2c4bdc6aa59ca6159286b89e6e38cf86325db29080191a13abad4a48d1254 not found: ID does not exist" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.976735 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-config-data\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.976841 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clvqb\" (UniqueName: \"kubernetes.io/projected/e9b54cbb-437d-4016-9ef7-8966ce4b9035-kube-api-access-clvqb\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.976862 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.980868 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.990553 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-config-data\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:00 crc kubenswrapper[4829]: I1002 07:37:00.996175 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clvqb\" (UniqueName: \"kubernetes.io/projected/e9b54cbb-437d-4016-9ef7-8966ce4b9035-kube-api-access-clvqb\") pod \"nova-scheduler-0\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.100584 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.100722 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.113550 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.137033 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.145898 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.149704 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.151929 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.160359 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.181324 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.181429 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-config-data\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.181658 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t66vr\" (UniqueName: \"kubernetes.io/projected/b380e2f7-3db9-439f-9282-fb93965b8cca-kube-api-access-t66vr\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.181808 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b380e2f7-3db9-439f-9282-fb93965b8cca-logs\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.217026 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.284807 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-config-data\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.284921 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t66vr\" (UniqueName: \"kubernetes.io/projected/b380e2f7-3db9-439f-9282-fb93965b8cca-kube-api-access-t66vr\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.284962 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b380e2f7-3db9-439f-9282-fb93965b8cca-logs\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.285129 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.285723 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b380e2f7-3db9-439f-9282-fb93965b8cca-logs\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.288716 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-config-data\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.290213 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.303405 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t66vr\" (UniqueName: \"kubernetes.io/projected/b380e2f7-3db9-439f-9282-fb93965b8cca-kube-api-access-t66vr\") pod \"nova-api-0\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.479466 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096" path="/var/lib/kubelet/pods/8e9679a7-bc7d-4ae0-8e8a-9fb59d13e096/volumes" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.480693 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="958f85b0-6cda-4f95-b713-4b0b59237d24" path="/var/lib/kubelet/pods/958f85b0-6cda-4f95-b713-4b0b59237d24/volumes" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.480932 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.743963 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.788078 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9b54cbb-437d-4016-9ef7-8966ce4b9035","Type":"ContainerStarted","Data":"31e42b7b1c820e452c3d9dd9a0bd80dfae69c0f98e508690a66c759c5196b970"} Oct 02 07:37:01 crc kubenswrapper[4829]: I1002 07:37:01.981191 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:02 crc kubenswrapper[4829]: I1002 07:37:02.799391 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9b54cbb-437d-4016-9ef7-8966ce4b9035","Type":"ContainerStarted","Data":"5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995"} Oct 02 07:37:02 crc kubenswrapper[4829]: I1002 07:37:02.803630 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b380e2f7-3db9-439f-9282-fb93965b8cca","Type":"ContainerStarted","Data":"c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878"} Oct 02 07:37:02 crc kubenswrapper[4829]: I1002 07:37:02.803667 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b380e2f7-3db9-439f-9282-fb93965b8cca","Type":"ContainerStarted","Data":"fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f"} Oct 02 07:37:02 crc kubenswrapper[4829]: I1002 07:37:02.803681 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b380e2f7-3db9-439f-9282-fb93965b8cca","Type":"ContainerStarted","Data":"44ff2589659ccb1edafa30bb9853bcb3fabfc1d4c3eae6028bb00080c1c16ee0"} Oct 02 07:37:02 crc kubenswrapper[4829]: I1002 07:37:02.824298 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.824203573 podStartE2EDuration="2.824203573s" podCreationTimestamp="2025-10-02 07:37:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:02.820413521 +0000 UTC m=+1214.160061926" watchObservedRunningTime="2025-10-02 07:37:02.824203573 +0000 UTC m=+1214.163851978" Oct 02 07:37:02 crc kubenswrapper[4829]: I1002 07:37:02.855203 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.855186327 podStartE2EDuration="1.855186327s" podCreationTimestamp="2025-10-02 07:37:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:02.852472069 +0000 UTC m=+1214.192120474" watchObservedRunningTime="2025-10-02 07:37:02.855186327 +0000 UTC m=+1214.194834732" Oct 02 07:37:03 crc kubenswrapper[4829]: I1002 07:37:03.816879 4829 generic.go:334] "Generic (PLEG): container finished" podID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerID="b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72" exitCode=0 Oct 02 07:37:03 crc kubenswrapper[4829]: I1002 07:37:03.816925 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerDied","Data":"b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72"} Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.023493 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036089 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-log-httpd\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036188 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-combined-ca-bundle\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036214 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-scripts\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036261 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-sg-core-conf-yaml\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036308 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-config-data\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036401 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-run-httpd\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.036446 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjln9\" (UniqueName: \"kubernetes.io/projected/a5d4e255-920f-4f35-89f6-08bc7355ff03-kube-api-access-bjln9\") pod \"a5d4e255-920f-4f35-89f6-08bc7355ff03\" (UID: \"a5d4e255-920f-4f35-89f6-08bc7355ff03\") " Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.037721 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.038056 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.043484 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5d4e255-920f-4f35-89f6-08bc7355ff03-kube-api-access-bjln9" (OuterVolumeSpecName: "kube-api-access-bjln9") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "kube-api-access-bjln9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.044430 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-scripts" (OuterVolumeSpecName: "scripts") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.133344 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.138456 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjln9\" (UniqueName: \"kubernetes.io/projected/a5d4e255-920f-4f35-89f6-08bc7355ff03-kube-api-access-bjln9\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.138508 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.138525 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.138539 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.138550 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a5d4e255-920f-4f35-89f6-08bc7355ff03-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.160629 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.210898 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-config-data" (OuterVolumeSpecName: "config-data") pod "a5d4e255-920f-4f35-89f6-08bc7355ff03" (UID: "a5d4e255-920f-4f35-89f6-08bc7355ff03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.240365 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.240402 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5d4e255-920f-4f35-89f6-08bc7355ff03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:04 crc kubenswrapper[4829]: E1002 07:37:04.341037 4829 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/31e029bcffc7909c625866c6f3e52d1ca32a55b3d2616bbff192101b23f96030/diff" to get inode usage: stat /var/lib/containers/storage/overlay/31e029bcffc7909c625866c6f3e52d1ca32a55b3d2616bbff192101b23f96030/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_kube-state-metrics-0_088860ca-f3ac-4f1b-86ae-13a3a13a8ba3/kube-state-metrics/0.log" to get inode usage: stat /var/log/pods/openstack_kube-state-metrics-0_088860ca-f3ac-4f1b-86ae-13a3a13a8ba3/kube-state-metrics/0.log: no such file or directory Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.828892 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a5d4e255-920f-4f35-89f6-08bc7355ff03","Type":"ContainerDied","Data":"425329d7716e3b0a3452c719f483c89ff6aad9bd0b818d8e00a8dee558702188"} Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.829288 4829 scope.go:117] "RemoveContainer" containerID="541258c680fb40bdf3dfe44364e81d92bf3100c73c2bdbf3cc519b48c3673666" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.828974 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.851051 4829 scope.go:117] "RemoveContainer" containerID="4fbb90c51caee159cef1056dd8484fa71b4577db9ac880c9dc1c443aa3abb4f5" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.881774 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.901548 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.905458 4829 scope.go:117] "RemoveContainer" containerID="b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.913611 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:04 crc kubenswrapper[4829]: E1002 07:37:04.914206 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="proxy-httpd" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914237 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="proxy-httpd" Oct 02 07:37:04 crc kubenswrapper[4829]: E1002 07:37:04.914254 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-central-agent" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914262 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-central-agent" Oct 02 07:37:04 crc kubenswrapper[4829]: E1002 07:37:04.914299 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-notification-agent" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914307 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-notification-agent" Oct 02 07:37:04 crc kubenswrapper[4829]: E1002 07:37:04.914318 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="sg-core" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914324 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="sg-core" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914545 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="proxy-httpd" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914557 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-notification-agent" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914572 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="sg-core" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.914587 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" containerName="ceilometer-central-agent" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.916567 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.921940 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.922211 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.922394 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.924100 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.931961 4829 scope.go:117] "RemoveContainer" containerID="13f6ccd385bcbd94cc65b2964ca612bf866091f1e4f8a9526fe7b83dec51ceb4" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.957945 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958022 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958050 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-log-httpd\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958099 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-run-httpd\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958131 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kph4q\" (UniqueName: \"kubernetes.io/projected/3132be3b-e406-4cf9-bca3-3377769b89f3-kube-api-access-kph4q\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958192 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-config-data\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958217 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-scripts\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:04 crc kubenswrapper[4829]: I1002 07:37:04.958309 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.004893 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.059929 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.059974 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-log-httpd\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.060026 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-run-httpd\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.060061 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kph4q\" (UniqueName: \"kubernetes.io/projected/3132be3b-e406-4cf9-bca3-3377769b89f3-kube-api-access-kph4q\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.060140 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-config-data\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.060165 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-scripts\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.060203 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.060263 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.062454 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-log-httpd\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.062569 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-run-httpd\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.067021 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.067059 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.067182 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-scripts\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.067339 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.067867 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-config-data\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.081539 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kph4q\" (UniqueName: \"kubernetes.io/projected/3132be3b-e406-4cf9-bca3-3377769b89f3-kube-api-access-kph4q\") pod \"ceilometer-0\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.235507 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.491000 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5d4e255-920f-4f35-89f6-08bc7355ff03" path="/var/lib/kubelet/pods/a5d4e255-920f-4f35-89f6-08bc7355ff03/volumes" Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.769968 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:05 crc kubenswrapper[4829]: I1002 07:37:05.846203 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerStarted","Data":"529cb2ca58b4a88a0c61486b0a17195e472809ed7ae612bd587c0ff29e5295b6"} Oct 02 07:37:06 crc kubenswrapper[4829]: I1002 07:37:06.081872 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 07:37:06 crc kubenswrapper[4829]: I1002 07:37:06.099552 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:37:06 crc kubenswrapper[4829]: I1002 07:37:06.099614 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:37:06 crc kubenswrapper[4829]: I1002 07:37:06.217301 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 07:37:06 crc kubenswrapper[4829]: I1002 07:37:06.857977 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerStarted","Data":"f87396dbec5f538ed77898ddb083401c14d7c10c66737693381003dee1dee9c3"} Oct 02 07:37:07 crc kubenswrapper[4829]: I1002 07:37:07.115393 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:07 crc kubenswrapper[4829]: I1002 07:37:07.115415 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:07 crc kubenswrapper[4829]: I1002 07:37:07.869490 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerStarted","Data":"6623ba5dc0516d6cb906e3297690c0b4ff614af92ec43a18c11620b5ba66be18"} Oct 02 07:37:08 crc kubenswrapper[4829]: I1002 07:37:08.880666 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerStarted","Data":"0f0e4bea90e8ebb06a39c8655b2a1e769b32928bd8fbaa2f29be0ce8ecce229e"} Oct 02 07:37:09 crc kubenswrapper[4829]: I1002 07:37:09.891035 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerStarted","Data":"9ec487a0c6d561dae1d7a1d985aaa20197319dec076f9d1944a1b6a07504dcd9"} Oct 02 07:37:09 crc kubenswrapper[4829]: I1002 07:37:09.892190 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:37:09 crc kubenswrapper[4829]: I1002 07:37:09.921855 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.241272136 podStartE2EDuration="5.921836971s" podCreationTimestamp="2025-10-02 07:37:04 +0000 UTC" firstStartedPulling="2025-10-02 07:37:05.779663942 +0000 UTC m=+1217.119312347" lastFinishedPulling="2025-10-02 07:37:09.460228767 +0000 UTC m=+1220.799877182" observedRunningTime="2025-10-02 07:37:09.917474699 +0000 UTC m=+1221.257123104" watchObservedRunningTime="2025-10-02 07:37:09.921836971 +0000 UTC m=+1221.261485376" Oct 02 07:37:11 crc kubenswrapper[4829]: I1002 07:37:11.217557 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 07:37:11 crc kubenswrapper[4829]: I1002 07:37:11.251788 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 07:37:11 crc kubenswrapper[4829]: I1002 07:37:11.482434 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:37:11 crc kubenswrapper[4829]: I1002 07:37:11.482492 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:37:11 crc kubenswrapper[4829]: I1002 07:37:11.933540 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 07:37:12 crc kubenswrapper[4829]: I1002 07:37:12.566415 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.213:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:12 crc kubenswrapper[4829]: I1002 07:37:12.566986 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.213:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:16 crc kubenswrapper[4829]: I1002 07:37:16.107023 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:37:16 crc kubenswrapper[4829]: I1002 07:37:16.111175 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:37:16 crc kubenswrapper[4829]: I1002 07:37:16.116193 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:37:16 crc kubenswrapper[4829]: I1002 07:37:16.975639 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:37:18 crc kubenswrapper[4829]: W1002 07:37:18.457831 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19bfc874_e2e6_462e_a0da_15fe04932f9e.slice/crio-956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d.scope WatchSource:0}: Error finding container 956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d: Status 404 returned error can't find the container with id 956df9db7ae67d5f4bc9dda40a1273108162a989c72562878e64bfbee364bf8d Oct 02 07:37:18 crc kubenswrapper[4829]: W1002 07:37:18.463733 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19bfc874_e2e6_462e_a0da_15fe04932f9e.slice/crio-f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385.scope WatchSource:0}: Error finding container f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385: Status 404 returned error can't find the container with id f06a61d48b62deac6a1bbf28d3babfa0744e2a55525dfe01a2acb0d7403e2385 Oct 02 07:37:18 crc kubenswrapper[4829]: E1002 07:37:18.783148 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e9679a7_bc7d_4ae0_8e8a_9fb59d13e096.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5d4e255_920f_4f35_89f6_08bc7355ff03.slice/crio-b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5d4e255_920f_4f35_89f6_08bc7355ff03.slice/crio-conmon-b86e70a61733c1f3a8400ecde30ce2b42d3a4c45daec89a8b53b733f39b91a72.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5d4e255_920f_4f35_89f6_08bc7355ff03.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5d4e255_920f_4f35_89f6_08bc7355ff03.slice/crio-425329d7716e3b0a3452c719f483c89ff6aad9bd0b818d8e00a8dee558702188\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc720e5ad_6e31_43ce_b7c1_a0b590e58a90.slice/crio-conmon-94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc720e5ad_6e31_43ce_b7c1_a0b590e58a90.slice/crio-94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e9679a7_bc7d_4ae0_8e8a_9fb59d13e096.slice/crio-1ac4e634d1f0b2b280f783adc2078759d9018f58e4c33a2968528868c76da78d\": RecentStats: unable to find data in memory cache]" Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.848434 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.960759 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b28nk\" (UniqueName: \"kubernetes.io/projected/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-kube-api-access-b28nk\") pod \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.961090 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-combined-ca-bundle\") pod \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.961134 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-config-data\") pod \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\" (UID: \"c720e5ad-6e31-43ce-b7c1-a0b590e58a90\") " Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.967356 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-kube-api-access-b28nk" (OuterVolumeSpecName: "kube-api-access-b28nk") pod "c720e5ad-6e31-43ce-b7c1-a0b590e58a90" (UID: "c720e5ad-6e31-43ce-b7c1-a0b590e58a90"). InnerVolumeSpecName "kube-api-access-b28nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.986881 4829 generic.go:334] "Generic (PLEG): container finished" podID="c720e5ad-6e31-43ce-b7c1-a0b590e58a90" containerID="94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661" exitCode=137 Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.987692 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.987687 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c720e5ad-6e31-43ce-b7c1-a0b590e58a90","Type":"ContainerDied","Data":"94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661"} Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.987848 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c720e5ad-6e31-43ce-b7c1-a0b590e58a90","Type":"ContainerDied","Data":"92bcead0765f53d51cb35e17aa73955f356e91231b50b72b68b939e735b0f16b"} Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.987873 4829 scope.go:117] "RemoveContainer" containerID="94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661" Oct 02 07:37:18 crc kubenswrapper[4829]: I1002 07:37:18.990806 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-config-data" (OuterVolumeSpecName: "config-data") pod "c720e5ad-6e31-43ce-b7c1-a0b590e58a90" (UID: "c720e5ad-6e31-43ce-b7c1-a0b590e58a90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.011429 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c720e5ad-6e31-43ce-b7c1-a0b590e58a90" (UID: "c720e5ad-6e31-43ce-b7c1-a0b590e58a90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.063356 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.063390 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.063402 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b28nk\" (UniqueName: \"kubernetes.io/projected/c720e5ad-6e31-43ce-b7c1-a0b590e58a90-kube-api-access-b28nk\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.120819 4829 scope.go:117] "RemoveContainer" containerID="94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661" Oct 02 07:37:19 crc kubenswrapper[4829]: E1002 07:37:19.121664 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661\": container with ID starting with 94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661 not found: ID does not exist" containerID="94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.121739 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661"} err="failed to get container status \"94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661\": rpc error: code = NotFound desc = could not find container \"94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661\": container with ID starting with 94e91fd253c950f88f137f128e91985c10e8f1d6593cc4ee7d2f5ba286cc3661 not found: ID does not exist" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.336740 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.357998 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.372125 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:37:19 crc kubenswrapper[4829]: E1002 07:37:19.372774 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c720e5ad-6e31-43ce-b7c1-a0b590e58a90" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.372838 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c720e5ad-6e31-43ce-b7c1-a0b590e58a90" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.373306 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c720e5ad-6e31-43ce-b7c1-a0b590e58a90" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.374771 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.378845 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.379818 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.382888 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.386465 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.471960 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.472034 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.472110 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxw78\" (UniqueName: \"kubernetes.io/projected/23a2af5c-929f-41a7-8a96-d42f5902f200-kube-api-access-cxw78\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.472130 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.472517 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.476019 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c720e5ad-6e31-43ce-b7c1-a0b590e58a90" path="/var/lib/kubelet/pods/c720e5ad-6e31-43ce-b7c1-a0b590e58a90/volumes" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.575155 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.575297 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.575419 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxw78\" (UniqueName: \"kubernetes.io/projected/23a2af5c-929f-41a7-8a96-d42f5902f200-kube-api-access-cxw78\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.575447 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.575581 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.581016 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.581560 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.582121 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.583797 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/23a2af5c-929f-41a7-8a96-d42f5902f200-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.593857 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxw78\" (UniqueName: \"kubernetes.io/projected/23a2af5c-929f-41a7-8a96-d42f5902f200-kube-api-access-cxw78\") pod \"nova-cell1-novncproxy-0\" (UID: \"23a2af5c-929f-41a7-8a96-d42f5902f200\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:19 crc kubenswrapper[4829]: I1002 07:37:19.706098 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:20 crc kubenswrapper[4829]: W1002 07:37:20.148866 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23a2af5c_929f_41a7_8a96_d42f5902f200.slice/crio-217e44feb88dbcc83a8006b48ed6f7043c5278853dbac5ba61c50140416b1554 WatchSource:0}: Error finding container 217e44feb88dbcc83a8006b48ed6f7043c5278853dbac5ba61c50140416b1554: Status 404 returned error can't find the container with id 217e44feb88dbcc83a8006b48ed6f7043c5278853dbac5ba61c50140416b1554 Oct 02 07:37:20 crc kubenswrapper[4829]: I1002 07:37:20.149829 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.012219 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"23a2af5c-929f-41a7-8a96-d42f5902f200","Type":"ContainerStarted","Data":"1073cd1b923a3c66ae3b6f0614974e840973074245d5e51e7eb3b59f6f58ea37"} Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.012802 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"23a2af5c-929f-41a7-8a96-d42f5902f200","Type":"ContainerStarted","Data":"217e44feb88dbcc83a8006b48ed6f7043c5278853dbac5ba61c50140416b1554"} Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.051112 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.051083403 podStartE2EDuration="2.051083403s" podCreationTimestamp="2025-10-02 07:37:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:21.030684332 +0000 UTC m=+1232.370332777" watchObservedRunningTime="2025-10-02 07:37:21.051083403 +0000 UTC m=+1232.390731848" Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.487036 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.488173 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.489352 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:37:21 crc kubenswrapper[4829]: I1002 07:37:21.491458 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.025732 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.032598 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.250348 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6wqhd"] Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.252614 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.281298 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6wqhd"] Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.434158 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-config\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.434209 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.434263 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.434334 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2mz\" (UniqueName: \"kubernetes.io/projected/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-kube-api-access-jl2mz\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.434379 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.434423 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.536209 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-config\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.536273 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.536299 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.536351 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2mz\" (UniqueName: \"kubernetes.io/projected/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-kube-api-access-jl2mz\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.536401 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.536436 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.537358 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.537844 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-config\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.538344 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.538868 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.539936 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.563418 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2mz\" (UniqueName: \"kubernetes.io/projected/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-kube-api-access-jl2mz\") pod \"dnsmasq-dns-89c5cd4d5-6wqhd\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:22 crc kubenswrapper[4829]: I1002 07:37:22.578735 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:23 crc kubenswrapper[4829]: I1002 07:37:23.110374 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6wqhd"] Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.054469 4829 generic.go:334] "Generic (PLEG): container finished" podID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerID="1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204" exitCode=0 Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.056614 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" event={"ID":"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c","Type":"ContainerDied","Data":"1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204"} Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.056649 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" event={"ID":"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c","Type":"ContainerStarted","Data":"1d91e88f78528cadd778e5e081befb332a44d930fe55bf5f60acb2e9f7cc3658"} Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.355678 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.356290 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-central-agent" containerID="cri-o://f87396dbec5f538ed77898ddb083401c14d7c10c66737693381003dee1dee9c3" gracePeriod=30 Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.356864 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="proxy-httpd" containerID="cri-o://9ec487a0c6d561dae1d7a1d985aaa20197319dec076f9d1944a1b6a07504dcd9" gracePeriod=30 Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.357069 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-notification-agent" containerID="cri-o://6623ba5dc0516d6cb906e3297690c0b4ff614af92ec43a18c11620b5ba66be18" gracePeriod=30 Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.357114 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="sg-core" containerID="cri-o://0f0e4bea90e8ebb06a39c8655b2a1e769b32928bd8fbaa2f29be0ce8ecce229e" gracePeriod=30 Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.369891 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.214:3000/\": read tcp 10.217.0.2:39806->10.217.0.214:3000: read: connection reset by peer" Oct 02 07:37:24 crc kubenswrapper[4829]: I1002 07:37:24.706179 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.064656 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" event={"ID":"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c","Type":"ContainerStarted","Data":"b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734"} Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.064962 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.067493 4829 generic.go:334] "Generic (PLEG): container finished" podID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerID="9ec487a0c6d561dae1d7a1d985aaa20197319dec076f9d1944a1b6a07504dcd9" exitCode=0 Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.067516 4829 generic.go:334] "Generic (PLEG): container finished" podID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerID="0f0e4bea90e8ebb06a39c8655b2a1e769b32928bd8fbaa2f29be0ce8ecce229e" exitCode=2 Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.067523 4829 generic.go:334] "Generic (PLEG): container finished" podID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerID="f87396dbec5f538ed77898ddb083401c14d7c10c66737693381003dee1dee9c3" exitCode=0 Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.067537 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerDied","Data":"9ec487a0c6d561dae1d7a1d985aaa20197319dec076f9d1944a1b6a07504dcd9"} Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.067552 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerDied","Data":"0f0e4bea90e8ebb06a39c8655b2a1e769b32928bd8fbaa2f29be0ce8ecce229e"} Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.067561 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerDied","Data":"f87396dbec5f538ed77898ddb083401c14d7c10c66737693381003dee1dee9c3"} Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.445019 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" podStartSLOduration=3.444989793 podStartE2EDuration="3.444989793s" podCreationTimestamp="2025-10-02 07:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:25.085402956 +0000 UTC m=+1236.425051361" watchObservedRunningTime="2025-10-02 07:37:25.444989793 +0000 UTC m=+1236.784638198" Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.445414 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.445660 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-log" containerID="cri-o://fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f" gracePeriod=30 Oct 02 07:37:25 crc kubenswrapper[4829]: I1002 07:37:25.446279 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-api" containerID="cri-o://c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878" gracePeriod=30 Oct 02 07:37:26 crc kubenswrapper[4829]: I1002 07:37:26.080850 4829 generic.go:334] "Generic (PLEG): container finished" podID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerID="fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f" exitCode=143 Oct 02 07:37:26 crc kubenswrapper[4829]: I1002 07:37:26.080945 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b380e2f7-3db9-439f-9282-fb93965b8cca","Type":"ContainerDied","Data":"fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f"} Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.106381 4829 generic.go:334] "Generic (PLEG): container finished" podID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerID="6623ba5dc0516d6cb906e3297690c0b4ff614af92ec43a18c11620b5ba66be18" exitCode=0 Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.106577 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerDied","Data":"6623ba5dc0516d6cb906e3297690c0b4ff614af92ec43a18c11620b5ba66be18"} Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.508093 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.664651 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-run-httpd\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665054 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-combined-ca-bundle\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.664952 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665116 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kph4q\" (UniqueName: \"kubernetes.io/projected/3132be3b-e406-4cf9-bca3-3377769b89f3-kube-api-access-kph4q\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665189 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-ceilometer-tls-certs\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665312 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-config-data\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665347 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-sg-core-conf-yaml\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665439 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-log-httpd\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665588 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-scripts\") pod \"3132be3b-e406-4cf9-bca3-3377769b89f3\" (UID: \"3132be3b-e406-4cf9-bca3-3377769b89f3\") " Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.665840 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.666344 4829 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.666404 4829 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3132be3b-e406-4cf9-bca3-3377769b89f3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.671359 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3132be3b-e406-4cf9-bca3-3377769b89f3-kube-api-access-kph4q" (OuterVolumeSpecName: "kube-api-access-kph4q") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "kube-api-access-kph4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.673946 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-scripts" (OuterVolumeSpecName: "scripts") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.698058 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.723693 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.750905 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.767586 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.767654 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kph4q\" (UniqueName: \"kubernetes.io/projected/3132be3b-e406-4cf9-bca3-3377769b89f3-kube-api-access-kph4q\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.767666 4829 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.767674 4829 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.767683 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.799912 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-config-data" (OuterVolumeSpecName: "config-data") pod "3132be3b-e406-4cf9-bca3-3377769b89f3" (UID: "3132be3b-e406-4cf9-bca3-3377769b89f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.870259 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3132be3b-e406-4cf9-bca3-3377769b89f3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:28 crc kubenswrapper[4829]: I1002 07:37:28.978736 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.121705 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3132be3b-e406-4cf9-bca3-3377769b89f3","Type":"ContainerDied","Data":"529cb2ca58b4a88a0c61486b0a17195e472809ed7ae612bd587c0ff29e5295b6"} Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.121722 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.122188 4829 scope.go:117] "RemoveContainer" containerID="9ec487a0c6d561dae1d7a1d985aaa20197319dec076f9d1944a1b6a07504dcd9" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.124748 4829 generic.go:334] "Generic (PLEG): container finished" podID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerID="c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878" exitCode=0 Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.124819 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b380e2f7-3db9-439f-9282-fb93965b8cca","Type":"ContainerDied","Data":"c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878"} Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.124908 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b380e2f7-3db9-439f-9282-fb93965b8cca","Type":"ContainerDied","Data":"44ff2589659ccb1edafa30bb9853bcb3fabfc1d4c3eae6028bb00080c1c16ee0"} Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.124845 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.171907 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.173940 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b380e2f7-3db9-439f-9282-fb93965b8cca-logs\") pod \"b380e2f7-3db9-439f-9282-fb93965b8cca\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.174015 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-config-data\") pod \"b380e2f7-3db9-439f-9282-fb93965b8cca\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.174114 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-combined-ca-bundle\") pod \"b380e2f7-3db9-439f-9282-fb93965b8cca\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.174150 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t66vr\" (UniqueName: \"kubernetes.io/projected/b380e2f7-3db9-439f-9282-fb93965b8cca-kube-api-access-t66vr\") pod \"b380e2f7-3db9-439f-9282-fb93965b8cca\" (UID: \"b380e2f7-3db9-439f-9282-fb93965b8cca\") " Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.174713 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b380e2f7-3db9-439f-9282-fb93965b8cca-logs" (OuterVolumeSpecName: "logs") pod "b380e2f7-3db9-439f-9282-fb93965b8cca" (UID: "b380e2f7-3db9-439f-9282-fb93965b8cca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.175484 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b380e2f7-3db9-439f-9282-fb93965b8cca-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.181636 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b380e2f7-3db9-439f-9282-fb93965b8cca-kube-api-access-t66vr" (OuterVolumeSpecName: "kube-api-access-t66vr") pod "b380e2f7-3db9-439f-9282-fb93965b8cca" (UID: "b380e2f7-3db9-439f-9282-fb93965b8cca"). InnerVolumeSpecName "kube-api-access-t66vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.189011 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.217710 4829 scope.go:117] "RemoveContainer" containerID="0f0e4bea90e8ebb06a39c8655b2a1e769b32928bd8fbaa2f29be0ce8ecce229e" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.221545 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-config-data" (OuterVolumeSpecName: "config-data") pod "b380e2f7-3db9-439f-9282-fb93965b8cca" (UID: "b380e2f7-3db9-439f-9282-fb93965b8cca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.221960 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.222387 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="proxy-httpd" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222404 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="proxy-httpd" Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.222417 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-central-agent" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222423 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-central-agent" Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.222450 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-log" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222456 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-log" Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.222465 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="sg-core" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222470 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="sg-core" Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.222482 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-api" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222488 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-api" Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.222508 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-notification-agent" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222514 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-notification-agent" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222823 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-log" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222840 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" containerName="nova-api-api" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222852 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-central-agent" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222865 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="sg-core" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222875 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="ceilometer-notification-agent" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.222883 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" containerName="proxy-httpd" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.226127 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.232569 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.236701 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.236902 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.237001 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277510 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277556 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-config-data\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277615 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277633 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-scripts\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277666 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjwr4\" (UniqueName: \"kubernetes.io/projected/409b81a0-0c77-4550-a88a-8489473fe397-kube-api-access-tjwr4\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277695 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277720 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/409b81a0-0c77-4550-a88a-8489473fe397-log-httpd\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277744 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/409b81a0-0c77-4550-a88a-8489473fe397-run-httpd\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277916 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.277930 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t66vr\" (UniqueName: \"kubernetes.io/projected/b380e2f7-3db9-439f-9282-fb93965b8cca-kube-api-access-t66vr\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.285855 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b380e2f7-3db9-439f-9282-fb93965b8cca" (UID: "b380e2f7-3db9-439f-9282-fb93965b8cca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.367978 4829 scope.go:117] "RemoveContainer" containerID="6623ba5dc0516d6cb906e3297690c0b4ff614af92ec43a18c11620b5ba66be18" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380061 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjwr4\" (UniqueName: \"kubernetes.io/projected/409b81a0-0c77-4550-a88a-8489473fe397-kube-api-access-tjwr4\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380138 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380175 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/409b81a0-0c77-4550-a88a-8489473fe397-log-httpd\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380194 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/409b81a0-0c77-4550-a88a-8489473fe397-run-httpd\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380275 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380294 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-config-data\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380350 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380374 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-scripts\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.380432 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b380e2f7-3db9-439f-9282-fb93965b8cca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.381180 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/409b81a0-0c77-4550-a88a-8489473fe397-run-httpd\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.381323 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/409b81a0-0c77-4550-a88a-8489473fe397-log-httpd\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.384179 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.384411 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.386086 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-config-data\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.386177 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.388287 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/409b81a0-0c77-4550-a88a-8489473fe397-scripts\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.401903 4829 scope.go:117] "RemoveContainer" containerID="f87396dbec5f538ed77898ddb083401c14d7c10c66737693381003dee1dee9c3" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.411639 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjwr4\" (UniqueName: \"kubernetes.io/projected/409b81a0-0c77-4550-a88a-8489473fe397-kube-api-access-tjwr4\") pod \"ceilometer-0\" (UID: \"409b81a0-0c77-4550-a88a-8489473fe397\") " pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.446569 4829 scope.go:117] "RemoveContainer" containerID="c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.475771 4829 scope.go:117] "RemoveContainer" containerID="fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.500705 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3132be3b-e406-4cf9-bca3-3377769b89f3" path="/var/lib/kubelet/pods/3132be3b-e406-4cf9-bca3-3377769b89f3/volumes" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.502150 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.502216 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.510368 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.512204 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.515875 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.515937 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.516169 4829 scope.go:117] "RemoveContainer" containerID="c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.516294 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.521415 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.522461 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878\": container with ID starting with c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878 not found: ID does not exist" containerID="c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.522600 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878"} err="failed to get container status \"c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878\": rpc error: code = NotFound desc = could not find container \"c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878\": container with ID starting with c1109930083ba42a478db230ad0dd475c53c48c71e232f9cd0d4c9ef4fe97878 not found: ID does not exist" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.522717 4829 scope.go:117] "RemoveContainer" containerID="fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f" Oct 02 07:37:29 crc kubenswrapper[4829]: E1002 07:37:29.523748 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f\": container with ID starting with fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f not found: ID does not exist" containerID="fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.523789 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f"} err="failed to get container status \"fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f\": rpc error: code = NotFound desc = could not find container \"fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f\": container with ID starting with fc5719ccb2a68ec98193bfb66c8854acc69fd231acb9edd8027bfcbd57dd199f not found: ID does not exist" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.651167 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.685497 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.685697 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-config-data\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.685777 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.685827 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-public-tls-certs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.685960 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7v2v\" (UniqueName: \"kubernetes.io/projected/07ddcd11-077f-4eb8-bb1f-9f032468aed2-kube-api-access-d7v2v\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.686140 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07ddcd11-077f-4eb8-bb1f-9f032468aed2-logs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.707288 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.730834 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.788174 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.788236 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-public-tls-certs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.788272 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7v2v\" (UniqueName: \"kubernetes.io/projected/07ddcd11-077f-4eb8-bb1f-9f032468aed2-kube-api-access-d7v2v\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.788306 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07ddcd11-077f-4eb8-bb1f-9f032468aed2-logs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.788344 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.788510 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-config-data\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.789464 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07ddcd11-077f-4eb8-bb1f-9f032468aed2-logs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.793107 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-config-data\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.793717 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-public-tls-certs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.794023 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.795608 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.812158 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7v2v\" (UniqueName: \"kubernetes.io/projected/07ddcd11-077f-4eb8-bb1f-9f032468aed2-kube-api-access-d7v2v\") pod \"nova-api-0\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " pod="openstack/nova-api-0" Oct 02 07:37:29 crc kubenswrapper[4829]: I1002 07:37:29.834579 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.117680 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.142212 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"409b81a0-0c77-4550-a88a-8489473fe397","Type":"ContainerStarted","Data":"c9b752057999a530887ec0b29847882695afa82dc04b20c4fa3fae4ce2e9b1ae"} Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.163097 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.314071 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.353054 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-8drj4"] Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.354255 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.357094 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.361854 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.364312 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8drj4"] Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.407481 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-config-data\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.407537 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-scripts\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.407843 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.407911 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cgqb\" (UniqueName: \"kubernetes.io/projected/d864f84e-5ba0-4fca-96e4-1fe63880960d-kube-api-access-4cgqb\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.509521 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.509566 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cgqb\" (UniqueName: \"kubernetes.io/projected/d864f84e-5ba0-4fca-96e4-1fe63880960d-kube-api-access-4cgqb\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.509655 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-config-data\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.509687 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-scripts\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.519915 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.520528 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-scripts\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.529017 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-config-data\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.530919 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cgqb\" (UniqueName: \"kubernetes.io/projected/d864f84e-5ba0-4fca-96e4-1fe63880960d-kube-api-access-4cgqb\") pod \"nova-cell1-cell-mapping-8drj4\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:30 crc kubenswrapper[4829]: I1002 07:37:30.817102 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:31 crc kubenswrapper[4829]: I1002 07:37:31.158007 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"409b81a0-0c77-4550-a88a-8489473fe397","Type":"ContainerStarted","Data":"a9de6c6f6a36e6ab9682dec9e67ce2df8de567ed9e1bff86751b52db15cb36a7"} Oct 02 07:37:31 crc kubenswrapper[4829]: I1002 07:37:31.162151 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"07ddcd11-077f-4eb8-bb1f-9f032468aed2","Type":"ContainerStarted","Data":"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702"} Oct 02 07:37:31 crc kubenswrapper[4829]: I1002 07:37:31.162180 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"07ddcd11-077f-4eb8-bb1f-9f032468aed2","Type":"ContainerStarted","Data":"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249"} Oct 02 07:37:31 crc kubenswrapper[4829]: I1002 07:37:31.162190 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"07ddcd11-077f-4eb8-bb1f-9f032468aed2","Type":"ContainerStarted","Data":"d717d59de7b1b8ee39fc7c1741fbefc81abb39bb2d0834a871061723b37c8943"} Oct 02 07:37:31 crc kubenswrapper[4829]: I1002 07:37:31.540246 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b380e2f7-3db9-439f-9282-fb93965b8cca" path="/var/lib/kubelet/pods/b380e2f7-3db9-439f-9282-fb93965b8cca/volumes" Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.143272 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.143252504 podStartE2EDuration="3.143252504s" podCreationTimestamp="2025-10-02 07:37:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:31.196965527 +0000 UTC m=+1242.536613932" watchObservedRunningTime="2025-10-02 07:37:32.143252504 +0000 UTC m=+1243.482900919" Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.149898 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-8drj4"] Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.171580 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8drj4" event={"ID":"d864f84e-5ba0-4fca-96e4-1fe63880960d","Type":"ContainerStarted","Data":"b24de7f88efdb3f14004ac432eef4eaefbafd6ce479a804669605af779d2f2c7"} Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.173951 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"409b81a0-0c77-4550-a88a-8489473fe397","Type":"ContainerStarted","Data":"635bbcf1874309ad7b615e07c19e1fec56c5345b3bd9f4e68db94ab9bf6dddd6"} Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.581085 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.695076 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gplfb"] Oct 02 07:37:32 crc kubenswrapper[4829]: I1002 07:37:32.695697 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="dnsmasq-dns" containerID="cri-o://c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970" gracePeriod=10 Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.137030 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.183568 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"409b81a0-0c77-4550-a88a-8489473fe397","Type":"ContainerStarted","Data":"af0f71ef8f51841f4c36e4e8b7d1501e01af3f7ef92b668283d838331f7850d9"} Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.185245 4829 generic.go:334] "Generic (PLEG): container finished" podID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerID="c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970" exitCode=0 Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.185292 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" event={"ID":"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121","Type":"ContainerDied","Data":"c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970"} Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.185320 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" event={"ID":"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121","Type":"ContainerDied","Data":"fc2d33fd17cfc1fdf2890079101e39c32176fd134d215951541a8bcf7be5a4d7"} Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.185341 4829 scope.go:117] "RemoveContainer" containerID="c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.185454 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.193547 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8drj4" event={"ID":"d864f84e-5ba0-4fca-96e4-1fe63880960d","Type":"ContainerStarted","Data":"6f98e26526394978d637124d05c083aa1dc0def656b4304adb9882cd3bd1afc3"} Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.210605 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-8drj4" podStartSLOduration=3.210585593 podStartE2EDuration="3.210585593s" podCreationTimestamp="2025-10-02 07:37:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:33.210437999 +0000 UTC m=+1244.550086404" watchObservedRunningTime="2025-10-02 07:37:33.210585593 +0000 UTC m=+1244.550234008" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.247755 4829 scope.go:117] "RemoveContainer" containerID="d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.282795 4829 scope.go:117] "RemoveContainer" containerID="c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970" Oct 02 07:37:33 crc kubenswrapper[4829]: E1002 07:37:33.283291 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970\": container with ID starting with c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970 not found: ID does not exist" containerID="c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.283325 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970"} err="failed to get container status \"c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970\": rpc error: code = NotFound desc = could not find container \"c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970\": container with ID starting with c73adc89132bc3a44046ad6bf6e3f0de3bda0508aee61b63081fb84a62639970 not found: ID does not exist" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.283351 4829 scope.go:117] "RemoveContainer" containerID="d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be" Oct 02 07:37:33 crc kubenswrapper[4829]: E1002 07:37:33.283723 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be\": container with ID starting with d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be not found: ID does not exist" containerID="d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.283754 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be"} err="failed to get container status \"d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be\": rpc error: code = NotFound desc = could not find container \"d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be\": container with ID starting with d14b731632d7316f30c27212207792f96176a233b1c84f4cfb36a7fff19657be not found: ID does not exist" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.293488 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-nb\") pod \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.293556 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-swift-storage-0\") pod \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.293611 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-svc\") pod \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.293647 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-sb\") pod \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.293830 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-config\") pod \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.293902 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd5t5\" (UniqueName: \"kubernetes.io/projected/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-kube-api-access-bd5t5\") pod \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\" (UID: \"9d06f0bc-84cd-477d-8fe4-0c2a28f7e121\") " Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.300382 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-kube-api-access-bd5t5" (OuterVolumeSpecName: "kube-api-access-bd5t5") pod "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" (UID: "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121"). InnerVolumeSpecName "kube-api-access-bd5t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.346575 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" (UID: "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.350661 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" (UID: "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.352643 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" (UID: "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.354972 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-config" (OuterVolumeSpecName: "config") pod "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" (UID: "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.356507 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" (UID: "9d06f0bc-84cd-477d-8fe4-0c2a28f7e121"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.395381 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.395412 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd5t5\" (UniqueName: \"kubernetes.io/projected/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-kube-api-access-bd5t5\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.395422 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.395431 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.395442 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.395449 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.513697 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gplfb"] Oct 02 07:37:33 crc kubenswrapper[4829]: I1002 07:37:33.522045 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-gplfb"] Oct 02 07:37:34 crc kubenswrapper[4829]: I1002 07:37:34.205124 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"409b81a0-0c77-4550-a88a-8489473fe397","Type":"ContainerStarted","Data":"b4d42b0aa89cf46461856c09758f3af083222cb74a5c91fabd802c91a3850174"} Oct 02 07:37:34 crc kubenswrapper[4829]: I1002 07:37:34.205498 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 07:37:34 crc kubenswrapper[4829]: I1002 07:37:34.240697 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.756566311 podStartE2EDuration="5.240679347s" podCreationTimestamp="2025-10-02 07:37:29 +0000 UTC" firstStartedPulling="2025-10-02 07:37:30.124005805 +0000 UTC m=+1241.463654210" lastFinishedPulling="2025-10-02 07:37:33.608118841 +0000 UTC m=+1244.947767246" observedRunningTime="2025-10-02 07:37:34.232084988 +0000 UTC m=+1245.571733393" watchObservedRunningTime="2025-10-02 07:37:34.240679347 +0000 UTC m=+1245.580327742" Oct 02 07:37:35 crc kubenswrapper[4829]: I1002 07:37:35.489497 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" path="/var/lib/kubelet/pods/9d06f0bc-84cd-477d-8fe4-0c2a28f7e121/volumes" Oct 02 07:37:37 crc kubenswrapper[4829]: I1002 07:37:37.262801 4829 generic.go:334] "Generic (PLEG): container finished" podID="d864f84e-5ba0-4fca-96e4-1fe63880960d" containerID="6f98e26526394978d637124d05c083aa1dc0def656b4304adb9882cd3bd1afc3" exitCode=0 Oct 02 07:37:37 crc kubenswrapper[4829]: I1002 07:37:37.262893 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8drj4" event={"ID":"d864f84e-5ba0-4fca-96e4-1fe63880960d","Type":"ContainerDied","Data":"6f98e26526394978d637124d05c083aa1dc0def656b4304adb9882cd3bd1afc3"} Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.074150 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-gplfb" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.206:5353: i/o timeout" Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.782951 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.902909 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cgqb\" (UniqueName: \"kubernetes.io/projected/d864f84e-5ba0-4fca-96e4-1fe63880960d-kube-api-access-4cgqb\") pod \"d864f84e-5ba0-4fca-96e4-1fe63880960d\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.903023 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-config-data\") pod \"d864f84e-5ba0-4fca-96e4-1fe63880960d\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.903132 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-combined-ca-bundle\") pod \"d864f84e-5ba0-4fca-96e4-1fe63880960d\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.903170 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-scripts\") pod \"d864f84e-5ba0-4fca-96e4-1fe63880960d\" (UID: \"d864f84e-5ba0-4fca-96e4-1fe63880960d\") " Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.909519 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d864f84e-5ba0-4fca-96e4-1fe63880960d-kube-api-access-4cgqb" (OuterVolumeSpecName: "kube-api-access-4cgqb") pod "d864f84e-5ba0-4fca-96e4-1fe63880960d" (UID: "d864f84e-5ba0-4fca-96e4-1fe63880960d"). InnerVolumeSpecName "kube-api-access-4cgqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.913397 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-scripts" (OuterVolumeSpecName: "scripts") pod "d864f84e-5ba0-4fca-96e4-1fe63880960d" (UID: "d864f84e-5ba0-4fca-96e4-1fe63880960d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.933307 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-config-data" (OuterVolumeSpecName: "config-data") pod "d864f84e-5ba0-4fca-96e4-1fe63880960d" (UID: "d864f84e-5ba0-4fca-96e4-1fe63880960d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:38 crc kubenswrapper[4829]: I1002 07:37:38.934656 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d864f84e-5ba0-4fca-96e4-1fe63880960d" (UID: "d864f84e-5ba0-4fca-96e4-1fe63880960d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.005489 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cgqb\" (UniqueName: \"kubernetes.io/projected/d864f84e-5ba0-4fca-96e4-1fe63880960d-kube-api-access-4cgqb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.005535 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.005548 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.005560 4829 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d864f84e-5ba0-4fca-96e4-1fe63880960d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.289426 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-8drj4" event={"ID":"d864f84e-5ba0-4fca-96e4-1fe63880960d","Type":"ContainerDied","Data":"b24de7f88efdb3f14004ac432eef4eaefbafd6ce479a804669605af779d2f2c7"} Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.289463 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b24de7f88efdb3f14004ac432eef4eaefbafd6ce479a804669605af779d2f2c7" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.289518 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-8drj4" Oct 02 07:37:39 crc kubenswrapper[4829]: E1002 07:37:39.340547 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd864f84e_5ba0_4fca_96e4_1fe63880960d.slice/crio-b24de7f88efdb3f14004ac432eef4eaefbafd6ce479a804669605af779d2f2c7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd864f84e_5ba0_4fca_96e4_1fe63880960d.slice\": RecentStats: unable to find data in memory cache]" Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.523149 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.523603 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-log" containerID="cri-o://ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249" gracePeriod=30 Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.523807 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-api" containerID="cri-o://6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702" gracePeriod=30 Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.529428 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.529627 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e9b54cbb-437d-4016-9ef7-8966ce4b9035" containerName="nova-scheduler-scheduler" containerID="cri-o://5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995" gracePeriod=30 Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.552736 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.552947 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-log" containerID="cri-o://1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3" gracePeriod=30 Oct 02 07:37:39 crc kubenswrapper[4829]: I1002 07:37:39.553343 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-metadata" containerID="cri-o://01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6" gracePeriod=30 Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.131081 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.226937 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7v2v\" (UniqueName: \"kubernetes.io/projected/07ddcd11-077f-4eb8-bb1f-9f032468aed2-kube-api-access-d7v2v\") pod \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.226995 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-combined-ca-bundle\") pod \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.227099 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07ddcd11-077f-4eb8-bb1f-9f032468aed2-logs\") pod \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.227125 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-config-data\") pod \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.227177 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-internal-tls-certs\") pod \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.227295 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-public-tls-certs\") pod \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\" (UID: \"07ddcd11-077f-4eb8-bb1f-9f032468aed2\") " Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.228768 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07ddcd11-077f-4eb8-bb1f-9f032468aed2-logs" (OuterVolumeSpecName: "logs") pod "07ddcd11-077f-4eb8-bb1f-9f032468aed2" (UID: "07ddcd11-077f-4eb8-bb1f-9f032468aed2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.234241 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ddcd11-077f-4eb8-bb1f-9f032468aed2-kube-api-access-d7v2v" (OuterVolumeSpecName: "kube-api-access-d7v2v") pod "07ddcd11-077f-4eb8-bb1f-9f032468aed2" (UID: "07ddcd11-077f-4eb8-bb1f-9f032468aed2"). InnerVolumeSpecName "kube-api-access-d7v2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.264485 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-config-data" (OuterVolumeSpecName: "config-data") pod "07ddcd11-077f-4eb8-bb1f-9f032468aed2" (UID: "07ddcd11-077f-4eb8-bb1f-9f032468aed2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.280476 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07ddcd11-077f-4eb8-bb1f-9f032468aed2" (UID: "07ddcd11-077f-4eb8-bb1f-9f032468aed2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.297933 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "07ddcd11-077f-4eb8-bb1f-9f032468aed2" (UID: "07ddcd11-077f-4eb8-bb1f-9f032468aed2"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.302394 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "07ddcd11-077f-4eb8-bb1f-9f032468aed2" (UID: "07ddcd11-077f-4eb8-bb1f-9f032468aed2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.308590 4829 generic.go:334] "Generic (PLEG): container finished" podID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerID="1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3" exitCode=143 Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.308766 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88ecb52c-ee97-45fc-9701-69d9d2abf67f","Type":"ContainerDied","Data":"1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3"} Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311387 4829 generic.go:334] "Generic (PLEG): container finished" podID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerID="6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702" exitCode=0 Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311463 4829 generic.go:334] "Generic (PLEG): container finished" podID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerID="ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249" exitCode=143 Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311521 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"07ddcd11-077f-4eb8-bb1f-9f032468aed2","Type":"ContainerDied","Data":"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702"} Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311580 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"07ddcd11-077f-4eb8-bb1f-9f032468aed2","Type":"ContainerDied","Data":"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249"} Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311631 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"07ddcd11-077f-4eb8-bb1f-9f032468aed2","Type":"ContainerDied","Data":"d717d59de7b1b8ee39fc7c1741fbefc81abb39bb2d0834a871061723b37c8943"} Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311690 4829 scope.go:117] "RemoveContainer" containerID="6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.311857 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.329337 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07ddcd11-077f-4eb8-bb1f-9f032468aed2-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.329380 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.329398 4829 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.329415 4829 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.329431 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7v2v\" (UniqueName: \"kubernetes.io/projected/07ddcd11-077f-4eb8-bb1f-9f032468aed2-kube-api-access-d7v2v\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.329446 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07ddcd11-077f-4eb8-bb1f-9f032468aed2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.347859 4829 scope.go:117] "RemoveContainer" containerID="ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.351423 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.357005 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.372466 4829 scope.go:117] "RemoveContainer" containerID="6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702" Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.373832 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702\": container with ID starting with 6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702 not found: ID does not exist" containerID="6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.373949 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702"} err="failed to get container status \"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702\": rpc error: code = NotFound desc = could not find container \"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702\": container with ID starting with 6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702 not found: ID does not exist" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.374017 4829 scope.go:117] "RemoveContainer" containerID="ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249" Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.376089 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249\": container with ID starting with ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249 not found: ID does not exist" containerID="ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.376182 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249"} err="failed to get container status \"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249\": rpc error: code = NotFound desc = could not find container \"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249\": container with ID starting with ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249 not found: ID does not exist" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.376331 4829 scope.go:117] "RemoveContainer" containerID="6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.376610 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702"} err="failed to get container status \"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702\": rpc error: code = NotFound desc = could not find container \"6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702\": container with ID starting with 6593f1c878682daf03bd1840d00df2c1f01405b941886888700dd43169f1f702 not found: ID does not exist" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.376675 4829 scope.go:117] "RemoveContainer" containerID="ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.376941 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249"} err="failed to get container status \"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249\": rpc error: code = NotFound desc = could not find container \"ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249\": container with ID starting with ec9ce7efbbc2f69e452e38020876d16b95aa449a2ea97ee1f13fa883c8686249 not found: ID does not exist" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.387105 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.387717 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-api" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.387778 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-api" Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.387857 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-log" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.387903 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-log" Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.387954 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="dnsmasq-dns" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388027 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="dnsmasq-dns" Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.388079 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d864f84e-5ba0-4fca-96e4-1fe63880960d" containerName="nova-manage" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388122 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d864f84e-5ba0-4fca-96e4-1fe63880960d" containerName="nova-manage" Oct 02 07:37:40 crc kubenswrapper[4829]: E1002 07:37:40.388173 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="init" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388215 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="init" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388468 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d06f0bc-84cd-477d-8fe4-0c2a28f7e121" containerName="dnsmasq-dns" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388625 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-api" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388683 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" containerName="nova-api-log" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.388733 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d864f84e-5ba0-4fca-96e4-1fe63880960d" containerName="nova-manage" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.389920 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.391874 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.392111 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.392118 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.394848 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.535250 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-config-data\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.535316 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-public-tls-certs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.535406 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.535431 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cswnv\" (UniqueName: \"kubernetes.io/projected/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-kube-api-access-cswnv\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.535495 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.535544 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-logs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.637513 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-config-data\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.637558 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-public-tls-certs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.637616 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.637640 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cswnv\" (UniqueName: \"kubernetes.io/projected/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-kube-api-access-cswnv\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.637671 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.637709 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-logs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.638103 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-logs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.641853 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-config-data\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.645427 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-public-tls-certs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.648154 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.665373 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.678407 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cswnv\" (UniqueName: \"kubernetes.io/projected/23175b6b-8b9a-43da-a7ba-d0b5cb04863a-kube-api-access-cswnv\") pod \"nova-api-0\" (UID: \"23175b6b-8b9a-43da-a7ba-d0b5cb04863a\") " pod="openstack/nova-api-0" Oct 02 07:37:40 crc kubenswrapper[4829]: I1002 07:37:40.719281 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.028196 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.165689 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-config-data\") pod \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.165760 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clvqb\" (UniqueName: \"kubernetes.io/projected/e9b54cbb-437d-4016-9ef7-8966ce4b9035-kube-api-access-clvqb\") pod \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.165958 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-combined-ca-bundle\") pod \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\" (UID: \"e9b54cbb-437d-4016-9ef7-8966ce4b9035\") " Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.174538 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9b54cbb-437d-4016-9ef7-8966ce4b9035-kube-api-access-clvqb" (OuterVolumeSpecName: "kube-api-access-clvqb") pod "e9b54cbb-437d-4016-9ef7-8966ce4b9035" (UID: "e9b54cbb-437d-4016-9ef7-8966ce4b9035"). InnerVolumeSpecName "kube-api-access-clvqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.222152 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9b54cbb-437d-4016-9ef7-8966ce4b9035" (UID: "e9b54cbb-437d-4016-9ef7-8966ce4b9035"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.230406 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-config-data" (OuterVolumeSpecName: "config-data") pod "e9b54cbb-437d-4016-9ef7-8966ce4b9035" (UID: "e9b54cbb-437d-4016-9ef7-8966ce4b9035"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.247905 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.271747 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.271785 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clvqb\" (UniqueName: \"kubernetes.io/projected/e9b54cbb-437d-4016-9ef7-8966ce4b9035-kube-api-access-clvqb\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.271800 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9b54cbb-437d-4016-9ef7-8966ce4b9035-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.322376 4829 generic.go:334] "Generic (PLEG): container finished" podID="e9b54cbb-437d-4016-9ef7-8966ce4b9035" containerID="5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995" exitCode=0 Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.322406 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.322435 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9b54cbb-437d-4016-9ef7-8966ce4b9035","Type":"ContainerDied","Data":"5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995"} Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.326825 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e9b54cbb-437d-4016-9ef7-8966ce4b9035","Type":"ContainerDied","Data":"31e42b7b1c820e452c3d9dd9a0bd80dfae69c0f98e508690a66c759c5196b970"} Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.326847 4829 scope.go:117] "RemoveContainer" containerID="5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.328261 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23175b6b-8b9a-43da-a7ba-d0b5cb04863a","Type":"ContainerStarted","Data":"305071b483a491a57799215106f1635440829ff8ae9678fd403d3949255b8c07"} Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.348772 4829 scope.go:117] "RemoveContainer" containerID="5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995" Oct 02 07:37:41 crc kubenswrapper[4829]: E1002 07:37:41.349128 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995\": container with ID starting with 5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995 not found: ID does not exist" containerID="5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.349170 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995"} err="failed to get container status \"5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995\": rpc error: code = NotFound desc = could not find container \"5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995\": container with ID starting with 5307e62c4a18db87faa030983f8a5e0e54cb9b37d3783b942bd6e6fd6b823995 not found: ID does not exist" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.369810 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.379268 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.391028 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:41 crc kubenswrapper[4829]: E1002 07:37:41.391447 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9b54cbb-437d-4016-9ef7-8966ce4b9035" containerName="nova-scheduler-scheduler" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.391468 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9b54cbb-437d-4016-9ef7-8966ce4b9035" containerName="nova-scheduler-scheduler" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.391706 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9b54cbb-437d-4016-9ef7-8966ce4b9035" containerName="nova-scheduler-scheduler" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.392427 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.394178 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.399970 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.470920 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ddcd11-077f-4eb8-bb1f-9f032468aed2" path="/var/lib/kubelet/pods/07ddcd11-077f-4eb8-bb1f-9f032468aed2/volumes" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.471529 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9b54cbb-437d-4016-9ef7-8966ce4b9035" path="/var/lib/kubelet/pods/e9b54cbb-437d-4016-9ef7-8966ce4b9035/volumes" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.475525 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f60623-b76f-43dc-9022-d62914e8950b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.475678 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlrmp\" (UniqueName: \"kubernetes.io/projected/52f60623-b76f-43dc-9022-d62914e8950b-kube-api-access-vlrmp\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.475738 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f60623-b76f-43dc-9022-d62914e8950b-config-data\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.577466 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlrmp\" (UniqueName: \"kubernetes.io/projected/52f60623-b76f-43dc-9022-d62914e8950b-kube-api-access-vlrmp\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.577529 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f60623-b76f-43dc-9022-d62914e8950b-config-data\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.577595 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f60623-b76f-43dc-9022-d62914e8950b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.581345 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52f60623-b76f-43dc-9022-d62914e8950b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.582019 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52f60623-b76f-43dc-9022-d62914e8950b-config-data\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.593307 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlrmp\" (UniqueName: \"kubernetes.io/projected/52f60623-b76f-43dc-9022-d62914e8950b-kube-api-access-vlrmp\") pod \"nova-scheduler-0\" (UID: \"52f60623-b76f-43dc-9022-d62914e8950b\") " pod="openstack/nova-scheduler-0" Oct 02 07:37:41 crc kubenswrapper[4829]: I1002 07:37:41.721188 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.215039 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.341836 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23175b6b-8b9a-43da-a7ba-d0b5cb04863a","Type":"ContainerStarted","Data":"ad83059f30a36b63062f1c747ffb8e8db3f0d5e2def287c1bd47fd13e169ab71"} Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.342203 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"23175b6b-8b9a-43da-a7ba-d0b5cb04863a","Type":"ContainerStarted","Data":"e3670964455255abf54e0428de11965ff5b44c1443eae2eb5b1fbbb2a6f13cf9"} Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.342953 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"52f60623-b76f-43dc-9022-d62914e8950b","Type":"ContainerStarted","Data":"5167601f2c2951cd3a42cad5d8104a7b21abd8601d01f3bd156be74c0401e3d3"} Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.367303 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.367281071 podStartE2EDuration="2.367281071s" podCreationTimestamp="2025-10-02 07:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:42.362200247 +0000 UTC m=+1253.701848652" watchObservedRunningTime="2025-10-02 07:37:42.367281071 +0000 UTC m=+1253.706929476" Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.710293 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": read tcp 10.217.0.2:34666->10.217.0.211:8775: read: connection reset by peer" Oct 02 07:37:42 crc kubenswrapper[4829]: I1002 07:37:42.710404 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": read tcp 10.217.0.2:34670->10.217.0.211:8775: read: connection reset by peer" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.237105 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.358959 4829 generic.go:334] "Generic (PLEG): container finished" podID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerID="01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6" exitCode=0 Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.359034 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88ecb52c-ee97-45fc-9701-69d9d2abf67f","Type":"ContainerDied","Data":"01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6"} Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.359064 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"88ecb52c-ee97-45fc-9701-69d9d2abf67f","Type":"ContainerDied","Data":"c1fd68a6acbd5b72b5b50a1b68677203e1519b2de488bc9297855795cb17ac72"} Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.359084 4829 scope.go:117] "RemoveContainer" containerID="01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.359252 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.370425 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"52f60623-b76f-43dc-9022-d62914e8950b","Type":"ContainerStarted","Data":"f6f658cded82f8e5f3cb5e6ff9e3c711f8cd5277e4e9c4fe3d92886217e00cab"} Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.387604 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.387578237 podStartE2EDuration="2.387578237s" podCreationTimestamp="2025-10-02 07:37:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:43.385058155 +0000 UTC m=+1254.724706570" watchObservedRunningTime="2025-10-02 07:37:43.387578237 +0000 UTC m=+1254.727226642" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.406272 4829 scope.go:117] "RemoveContainer" containerID="1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.437166 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ecb52c-ee97-45fc-9701-69d9d2abf67f-logs\") pod \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.437281 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-combined-ca-bundle\") pod \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.437347 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-config-data\") pod \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.437409 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-nova-metadata-tls-certs\") pod \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.437442 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66ddh\" (UniqueName: \"kubernetes.io/projected/88ecb52c-ee97-45fc-9701-69d9d2abf67f-kube-api-access-66ddh\") pod \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\" (UID: \"88ecb52c-ee97-45fc-9701-69d9d2abf67f\") " Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.438977 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88ecb52c-ee97-45fc-9701-69d9d2abf67f-logs" (OuterVolumeSpecName: "logs") pod "88ecb52c-ee97-45fc-9701-69d9d2abf67f" (UID: "88ecb52c-ee97-45fc-9701-69d9d2abf67f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.445730 4829 scope.go:117] "RemoveContainer" containerID="01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6" Oct 02 07:37:43 crc kubenswrapper[4829]: E1002 07:37:43.446374 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6\": container with ID starting with 01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6 not found: ID does not exist" containerID="01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.446434 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6"} err="failed to get container status \"01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6\": rpc error: code = NotFound desc = could not find container \"01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6\": container with ID starting with 01db88f6485bce97c2db8b9792012a1ce5a4b0e196369bf4ddbab79a074bc7b6 not found: ID does not exist" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.446467 4829 scope.go:117] "RemoveContainer" containerID="1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3" Oct 02 07:37:43 crc kubenswrapper[4829]: E1002 07:37:43.447383 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3\": container with ID starting with 1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3 not found: ID does not exist" containerID="1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.447442 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3"} err="failed to get container status \"1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3\": rpc error: code = NotFound desc = could not find container \"1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3\": container with ID starting with 1e262e19544a824acb32dd18ed570841d1b0fe2fbc4ee90412eb721116d48ca3 not found: ID does not exist" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.454657 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ecb52c-ee97-45fc-9701-69d9d2abf67f-kube-api-access-66ddh" (OuterVolumeSpecName: "kube-api-access-66ddh") pod "88ecb52c-ee97-45fc-9701-69d9d2abf67f" (UID: "88ecb52c-ee97-45fc-9701-69d9d2abf67f"). InnerVolumeSpecName "kube-api-access-66ddh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.477751 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88ecb52c-ee97-45fc-9701-69d9d2abf67f" (UID: "88ecb52c-ee97-45fc-9701-69d9d2abf67f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.480503 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-config-data" (OuterVolumeSpecName: "config-data") pod "88ecb52c-ee97-45fc-9701-69d9d2abf67f" (UID: "88ecb52c-ee97-45fc-9701-69d9d2abf67f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.499462 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "88ecb52c-ee97-45fc-9701-69d9d2abf67f" (UID: "88ecb52c-ee97-45fc-9701-69d9d2abf67f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.543165 4829 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88ecb52c-ee97-45fc-9701-69d9d2abf67f-logs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.543209 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.543257 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.543283 4829 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/88ecb52c-ee97-45fc-9701-69d9d2abf67f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.543310 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66ddh\" (UniqueName: \"kubernetes.io/projected/88ecb52c-ee97-45fc-9701-69d9d2abf67f-kube-api-access-66ddh\") on node \"crc\" DevicePath \"\"" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.700547 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.711879 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.725166 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:37:43 crc kubenswrapper[4829]: E1002 07:37:43.728260 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-log" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.728296 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-log" Oct 02 07:37:43 crc kubenswrapper[4829]: E1002 07:37:43.728313 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-metadata" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.728322 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-metadata" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.728598 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-log" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.728627 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" containerName="nova-metadata-metadata" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.730248 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.740243 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.740487 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.741513 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.766709 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-config-data\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.766754 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1298703b-f77d-4cf5-9537-8928493b4e6b-logs\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.766860 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5wwk\" (UniqueName: \"kubernetes.io/projected/1298703b-f77d-4cf5-9537-8928493b4e6b-kube-api-access-d5wwk\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.766893 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.766986 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.868640 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.868787 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-config-data\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.868835 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1298703b-f77d-4cf5-9537-8928493b4e6b-logs\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.868893 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5wwk\" (UniqueName: \"kubernetes.io/projected/1298703b-f77d-4cf5-9537-8928493b4e6b-kube-api-access-d5wwk\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.868939 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.869454 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1298703b-f77d-4cf5-9537-8928493b4e6b-logs\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.872655 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-config-data\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.873140 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.873289 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1298703b-f77d-4cf5-9537-8928493b4e6b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:43 crc kubenswrapper[4829]: I1002 07:37:43.890637 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5wwk\" (UniqueName: \"kubernetes.io/projected/1298703b-f77d-4cf5-9537-8928493b4e6b-kube-api-access-d5wwk\") pod \"nova-metadata-0\" (UID: \"1298703b-f77d-4cf5-9537-8928493b4e6b\") " pod="openstack/nova-metadata-0" Oct 02 07:37:44 crc kubenswrapper[4829]: I1002 07:37:44.066497 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 07:37:44 crc kubenswrapper[4829]: I1002 07:37:44.582135 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 07:37:44 crc kubenswrapper[4829]: W1002 07:37:44.600037 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1298703b_f77d_4cf5_9537_8928493b4e6b.slice/crio-5ad71215aae9149ac6714323bcddbe993e2cefec4af3d8ebcbc47a5b9d6f976b WatchSource:0}: Error finding container 5ad71215aae9149ac6714323bcddbe993e2cefec4af3d8ebcbc47a5b9d6f976b: Status 404 returned error can't find the container with id 5ad71215aae9149ac6714323bcddbe993e2cefec4af3d8ebcbc47a5b9d6f976b Oct 02 07:37:45 crc kubenswrapper[4829]: I1002 07:37:45.393591 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1298703b-f77d-4cf5-9537-8928493b4e6b","Type":"ContainerStarted","Data":"8324363cd19326a37e872188d7e7ebcf5edb80d915ea58754cb894d65165c74b"} Oct 02 07:37:45 crc kubenswrapper[4829]: I1002 07:37:45.394199 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1298703b-f77d-4cf5-9537-8928493b4e6b","Type":"ContainerStarted","Data":"fe01a440c5c4fb05ee9bf9beedd2cc86b86deb990f0383e59f6adddd6295289e"} Oct 02 07:37:45 crc kubenswrapper[4829]: I1002 07:37:45.394343 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1298703b-f77d-4cf5-9537-8928493b4e6b","Type":"ContainerStarted","Data":"5ad71215aae9149ac6714323bcddbe993e2cefec4af3d8ebcbc47a5b9d6f976b"} Oct 02 07:37:45 crc kubenswrapper[4829]: I1002 07:37:45.412812 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.412793459 podStartE2EDuration="2.412793459s" podCreationTimestamp="2025-10-02 07:37:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:37:45.411817818 +0000 UTC m=+1256.751466233" watchObservedRunningTime="2025-10-02 07:37:45.412793459 +0000 UTC m=+1256.752441884" Oct 02 07:37:45 crc kubenswrapper[4829]: I1002 07:37:45.483128 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ecb52c-ee97-45fc-9701-69d9d2abf67f" path="/var/lib/kubelet/pods/88ecb52c-ee97-45fc-9701-69d9d2abf67f/volumes" Oct 02 07:37:46 crc kubenswrapper[4829]: I1002 07:37:46.721967 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 07:37:49 crc kubenswrapper[4829]: I1002 07:37:49.067339 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:37:49 crc kubenswrapper[4829]: I1002 07:37:49.067578 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 07:37:50 crc kubenswrapper[4829]: I1002 07:37:50.719863 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:37:50 crc kubenswrapper[4829]: I1002 07:37:50.720130 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 07:37:51 crc kubenswrapper[4829]: I1002 07:37:51.721368 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 07:37:51 crc kubenswrapper[4829]: I1002 07:37:51.734490 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23175b6b-8b9a-43da-a7ba-d0b5cb04863a" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:51 crc kubenswrapper[4829]: I1002 07:37:51.734509 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="23175b6b-8b9a-43da-a7ba-d0b5cb04863a" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.220:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:51 crc kubenswrapper[4829]: I1002 07:37:51.770034 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 07:37:52 crc kubenswrapper[4829]: I1002 07:37:52.542952 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 07:37:54 crc kubenswrapper[4829]: I1002 07:37:54.067400 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:37:54 crc kubenswrapper[4829]: I1002 07:37:54.067811 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 07:37:55 crc kubenswrapper[4829]: I1002 07:37:55.086449 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1298703b-f77d-4cf5-9537-8928493b4e6b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.222:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:55 crc kubenswrapper[4829]: I1002 07:37:55.086741 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1298703b-f77d-4cf5-9537-8928493b4e6b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.222:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 07:37:55 crc kubenswrapper[4829]: I1002 07:37:55.329295 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:37:55 crc kubenswrapper[4829]: I1002 07:37:55.329362 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:37:59 crc kubenswrapper[4829]: I1002 07:37:59.664547 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 07:38:00 crc kubenswrapper[4829]: I1002 07:38:00.726355 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:38:00 crc kubenswrapper[4829]: I1002 07:38:00.727063 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:38:00 crc kubenswrapper[4829]: I1002 07:38:00.727960 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 07:38:00 crc kubenswrapper[4829]: I1002 07:38:00.733209 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:38:01 crc kubenswrapper[4829]: I1002 07:38:01.598252 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 07:38:01 crc kubenswrapper[4829]: I1002 07:38:01.610036 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 07:38:04 crc kubenswrapper[4829]: I1002 07:38:04.072135 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:38:04 crc kubenswrapper[4829]: I1002 07:38:04.073104 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 07:38:04 crc kubenswrapper[4829]: I1002 07:38:04.080937 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:38:04 crc kubenswrapper[4829]: I1002 07:38:04.657849 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 07:38:13 crc kubenswrapper[4829]: I1002 07:38:13.982756 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:38:14 crc kubenswrapper[4829]: I1002 07:38:14.951589 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:38:18 crc kubenswrapper[4829]: I1002 07:38:18.013254 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="rabbitmq" containerID="cri-o://270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e" gracePeriod=604796 Oct 02 07:38:19 crc kubenswrapper[4829]: I1002 07:38:19.357844 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="rabbitmq" containerID="cri-o://47a462567327e8b2e3336281a0ec373ccb15e3712fdae76fb61751f95c28d012" gracePeriod=604796 Oct 02 07:38:21 crc kubenswrapper[4829]: I1002 07:38:21.075594 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.107:5671: connect: connection refused" Oct 02 07:38:21 crc kubenswrapper[4829]: I1002 07:38:21.402627 4829 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.108:5671: connect: connection refused" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.605809 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.676825 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.676913 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-tls\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.676947 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-config-data\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.676993 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs9k4\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-kube-api-access-rs9k4\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677044 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-server-conf\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677083 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-confd\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677142 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-erlang-cookie\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677241 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-plugins-conf\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677269 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-erlang-cookie-secret\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677313 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-plugins\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.677444 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-pod-info\") pod \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\" (UID: \"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2\") " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.678055 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.678097 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.678264 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.687909 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.692334 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-pod-info" (OuterVolumeSpecName: "pod-info") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.695938 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.696066 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-kube-api-access-rs9k4" (OuterVolumeSpecName: "kube-api-access-rs9k4") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "kube-api-access-rs9k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.706092 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.741316 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-config-data" (OuterVolumeSpecName: "config-data") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780006 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780088 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs9k4\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-kube-api-access-rs9k4\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780137 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780148 4829 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780157 4829 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780167 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780174 4829 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780194 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.780235 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.789835 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-server-conf" (OuterVolumeSpecName: "server-conf") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.801410 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.805079 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" (UID: "3c3f3f0f-d5d1-462c-9a48-3b25c840abb2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.871800 4829 generic.go:334] "Generic (PLEG): container finished" podID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerID="270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e" exitCode=0 Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.871840 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2","Type":"ContainerDied","Data":"270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e"} Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.871865 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3c3f3f0f-d5d1-462c-9a48-3b25c840abb2","Type":"ContainerDied","Data":"c1623bd0c6b05992aa4dfff70b38975e0735f2ca4df4b1a23138c7fb3220f8a7"} Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.871880 4829 scope.go:117] "RemoveContainer" containerID="270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.871984 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.889735 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.889947 4829 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.890034 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.913474 4829 scope.go:117] "RemoveContainer" containerID="3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.919140 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.941699 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.955680 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.955806 4829 scope.go:117] "RemoveContainer" containerID="270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e" Oct 02 07:38:24 crc kubenswrapper[4829]: E1002 07:38:24.956092 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="setup-container" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.956108 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="setup-container" Oct 02 07:38:24 crc kubenswrapper[4829]: E1002 07:38:24.956125 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="rabbitmq" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.956131 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="rabbitmq" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.956326 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" containerName="rabbitmq" Oct 02 07:38:24 crc kubenswrapper[4829]: E1002 07:38:24.956458 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e\": container with ID starting with 270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e not found: ID does not exist" containerID="270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.956498 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e"} err="failed to get container status \"270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e\": rpc error: code = NotFound desc = could not find container \"270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e\": container with ID starting with 270feaffa5effeaaefeca5d5e28bc8ef50e6f1babf99cd68ea5e2fa308445c2e not found: ID does not exist" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.956517 4829 scope.go:117] "RemoveContainer" containerID="3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7" Oct 02 07:38:24 crc kubenswrapper[4829]: E1002 07:38:24.956746 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7\": container with ID starting with 3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7 not found: ID does not exist" containerID="3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.956761 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7"} err="failed to get container status \"3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7\": rpc error: code = NotFound desc = could not find container \"3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7\": container with ID starting with 3623ca96f8f20b47acaf9242b40b29ab2bf41f41589c73feb1212e7de29875c7 not found: ID does not exist" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.957327 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.960278 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.960408 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.960509 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.960645 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-sqk68" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.960701 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.962867 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.966417 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 07:38:24 crc kubenswrapper[4829]: I1002 07:38:24.972149 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093443 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093703 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093745 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43522a1c-881d-4d33-91e3-3dea8c241d8c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093810 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hwhw\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-kube-api-access-9hwhw\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093832 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093853 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093885 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093911 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093926 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-config-data\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.093963 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.094019 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43522a1c-881d-4d33-91e3-3dea8c241d8c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196456 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43522a1c-881d-4d33-91e3-3dea8c241d8c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196599 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196635 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196720 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43522a1c-881d-4d33-91e3-3dea8c241d8c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196881 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hwhw\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-kube-api-access-9hwhw\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196919 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.196948 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.197014 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.197058 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.197076 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-config-data\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.197162 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.198112 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.198419 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.198166 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.198836 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-config-data\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.200258 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.200677 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/43522a1c-881d-4d33-91e3-3dea8c241d8c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.201690 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/43522a1c-881d-4d33-91e3-3dea8c241d8c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.202161 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.214936 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/43522a1c-881d-4d33-91e3-3dea8c241d8c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.219759 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.221259 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hwhw\" (UniqueName: \"kubernetes.io/projected/43522a1c-881d-4d33-91e3-3dea8c241d8c-kube-api-access-9hwhw\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.273846 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"43522a1c-881d-4d33-91e3-3dea8c241d8c\") " pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.329146 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.329215 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.475278 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3f3f0f-d5d1-462c-9a48-3b25c840abb2" path="/var/lib/kubelet/pods/3c3f3f0f-d5d1-462c-9a48-3b25c840abb2/volumes" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.576021 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.885429 4829 generic.go:334] "Generic (PLEG): container finished" podID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerID="47a462567327e8b2e3336281a0ec373ccb15e3712fdae76fb61751f95c28d012" exitCode=0 Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.885494 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bf3dfd-bae4-4af6-a338-fad9857bfd29","Type":"ContainerDied","Data":"47a462567327e8b2e3336281a0ec373ccb15e3712fdae76fb61751f95c28d012"} Oct 02 07:38:25 crc kubenswrapper[4829]: I1002 07:38:25.986583 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113157 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bf3dfd-bae4-4af6-a338-fad9857bfd29-erlang-cookie-secret\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113382 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-confd\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113479 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bf3dfd-bae4-4af6-a338-fad9857bfd29-pod-info\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113573 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7cbl\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-kube-api-access-x7cbl\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113680 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-tls\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113783 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-plugins\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113886 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.113993 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-config-data\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.114120 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-server-conf\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.114213 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-plugins-conf\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.114274 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.114443 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-erlang-cookie\") pod \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\" (UID: \"56bf3dfd-bae4-4af6-a338-fad9857bfd29\") " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.114898 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.114915 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.115269 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.118789 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.119152 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56bf3dfd-bae4-4af6-a338-fad9857bfd29-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.119551 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/56bf3dfd-bae4-4af6-a338-fad9857bfd29-pod-info" (OuterVolumeSpecName: "pod-info") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.119871 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-kube-api-access-x7cbl" (OuterVolumeSpecName: "kube-api-access-x7cbl") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "kube-api-access-x7cbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.120591 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.155375 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-config-data" (OuterVolumeSpecName: "config-data") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.169598 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-server-conf" (OuterVolumeSpecName: "server-conf") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.182520 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216305 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7cbl\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-kube-api-access-x7cbl\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216336 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216360 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216459 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216471 4829 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216479 4829 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/56bf3dfd-bae4-4af6-a338-fad9857bfd29-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216488 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216495 4829 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/56bf3dfd-bae4-4af6-a338-fad9857bfd29-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.216513 4829 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/56bf3dfd-bae4-4af6-a338-fad9857bfd29-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.240643 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "56bf3dfd-bae4-4af6-a338-fad9857bfd29" (UID: "56bf3dfd-bae4-4af6-a338-fad9857bfd29"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.244309 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.318053 4829 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/56bf3dfd-bae4-4af6-a338-fad9857bfd29-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.318089 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.903626 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"56bf3dfd-bae4-4af6-a338-fad9857bfd29","Type":"ContainerDied","Data":"862ff59d1bbc09414b573d7a80d32493b48ffe9b180cc0aa329338c6e63ae7ee"} Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.903679 4829 scope.go:117] "RemoveContainer" containerID="47a462567327e8b2e3336281a0ec373ccb15e3712fdae76fb61751f95c28d012" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.903844 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.907714 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43522a1c-881d-4d33-91e3-3dea8c241d8c","Type":"ContainerStarted","Data":"96a81cb59f2e1dbd6fc611262022311128fa6bf11159f34e2fb1b7296efe401e"} Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.955698 4829 scope.go:117] "RemoveContainer" containerID="fad8f75f06da64a1030f412dfcd4668bdcec0ea2d6b25739b69ddbef0345d357" Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.958326 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:38:26 crc kubenswrapper[4829]: I1002 07:38:26.970965 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.012947 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:38:27 crc kubenswrapper[4829]: E1002 07:38:27.013599 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="rabbitmq" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.013628 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="rabbitmq" Oct 02 07:38:27 crc kubenswrapper[4829]: E1002 07:38:27.013662 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="setup-container" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.013674 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="setup-container" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.014052 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" containerName="rabbitmq" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.015840 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.023660 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.023912 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.024038 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.024198 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.024359 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.024478 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-tdd4x" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.027952 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.031197 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132520 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132606 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132678 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132755 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132790 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132821 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132891 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132924 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbvbc\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-kube-api-access-mbvbc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.132980 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b66249bb-0358-455a-8ca1-7c9688cfe5f0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.133087 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.133160 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b66249bb-0358-455a-8ca1-7c9688cfe5f0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.234532 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.234974 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.235153 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.234972 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.235364 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236026 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236068 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbvbc\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-kube-api-access-mbvbc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236155 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b66249bb-0358-455a-8ca1-7c9688cfe5f0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236212 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236278 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b66249bb-0358-455a-8ca1-7c9688cfe5f0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236363 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.236426 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.237077 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.237675 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.238059 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.238092 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b66249bb-0358-455a-8ca1-7c9688cfe5f0-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.239657 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.293060 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b66249bb-0358-455a-8ca1-7c9688cfe5f0-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.293187 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b66249bb-0358-455a-8ca1-7c9688cfe5f0-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.293671 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.293847 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.295709 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbvbc\" (UniqueName: \"kubernetes.io/projected/b66249bb-0358-455a-8ca1-7c9688cfe5f0-kube-api-access-mbvbc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.425391 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b66249bb-0358-455a-8ca1-7c9688cfe5f0\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.474726 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56bf3dfd-bae4-4af6-a338-fad9857bfd29" path="/var/lib/kubelet/pods/56bf3dfd-bae4-4af6-a338-fad9857bfd29/volumes" Oct 02 07:38:27 crc kubenswrapper[4829]: I1002 07:38:27.657708 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:38:28 crc kubenswrapper[4829]: I1002 07:38:28.152707 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 07:38:28 crc kubenswrapper[4829]: W1002 07:38:28.155746 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb66249bb_0358_455a_8ca1_7c9688cfe5f0.slice/crio-65e3fcaca8ee18816d64a77f3bb2d25d58f04f0045fc9997ec039e16b34f1120 WatchSource:0}: Error finding container 65e3fcaca8ee18816d64a77f3bb2d25d58f04f0045fc9997ec039e16b34f1120: Status 404 returned error can't find the container with id 65e3fcaca8ee18816d64a77f3bb2d25d58f04f0045fc9997ec039e16b34f1120 Oct 02 07:38:28 crc kubenswrapper[4829]: I1002 07:38:28.948009 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66249bb-0358-455a-8ca1-7c9688cfe5f0","Type":"ContainerStarted","Data":"65e3fcaca8ee18816d64a77f3bb2d25d58f04f0045fc9997ec039e16b34f1120"} Oct 02 07:38:28 crc kubenswrapper[4829]: I1002 07:38:28.949502 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43522a1c-881d-4d33-91e3-3dea8c241d8c","Type":"ContainerStarted","Data":"5fb7cef4886633b52a13a40754ce132a0375ed51da9eb62e32d41135c0bd1d0f"} Oct 02 07:38:29 crc kubenswrapper[4829]: I1002 07:38:29.881481 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-c7fhz"] Oct 02 07:38:29 crc kubenswrapper[4829]: I1002 07:38:29.883374 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:29 crc kubenswrapper[4829]: I1002 07:38:29.885684 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 02 07:38:29 crc kubenswrapper[4829]: I1002 07:38:29.905570 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-c7fhz"] Oct 02 07:38:29 crc kubenswrapper[4829]: I1002 07:38:29.961763 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66249bb-0358-455a-8ca1-7c9688cfe5f0","Type":"ContainerStarted","Data":"d168b697068a6173b56c693ccb7c2971c88697528998bc48a515ed1e792e3142"} Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.003985 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-config\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.004468 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.004493 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.004593 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2vhl\" (UniqueName: \"kubernetes.io/projected/d9c5cc87-c132-43da-a319-1c65acf3c714-kube-api-access-k2vhl\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.004644 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.004671 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.004836 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106364 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2vhl\" (UniqueName: \"kubernetes.io/projected/d9c5cc87-c132-43da-a319-1c65acf3c714-kube-api-access-k2vhl\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106430 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106451 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106544 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106576 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106593 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-config\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.106611 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.107632 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.107680 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.107708 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.108339 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.108396 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-config\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.108460 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.138582 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2vhl\" (UniqueName: \"kubernetes.io/projected/d9c5cc87-c132-43da-a319-1c65acf3c714-kube-api-access-k2vhl\") pod \"dnsmasq-dns-79bd4cc8c9-c7fhz\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.206811 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:30 crc kubenswrapper[4829]: I1002 07:38:30.866209 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-c7fhz"] Oct 02 07:38:31 crc kubenswrapper[4829]: I1002 07:38:31.026402 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" event={"ID":"d9c5cc87-c132-43da-a319-1c65acf3c714","Type":"ContainerStarted","Data":"973fb12556b7ab018a2170eacc91f13e7a13f121cce77bd0ba42fed58b66a80c"} Oct 02 07:38:32 crc kubenswrapper[4829]: I1002 07:38:32.036340 4829 generic.go:334] "Generic (PLEG): container finished" podID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerID="a17d896a685a3e3ba84ea3fb874c02c80d8b3106858f148c4322939e9143e154" exitCode=0 Oct 02 07:38:32 crc kubenswrapper[4829]: I1002 07:38:32.036653 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" event={"ID":"d9c5cc87-c132-43da-a319-1c65acf3c714","Type":"ContainerDied","Data":"a17d896a685a3e3ba84ea3fb874c02c80d8b3106858f148c4322939e9143e154"} Oct 02 07:38:33 crc kubenswrapper[4829]: I1002 07:38:33.054715 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" event={"ID":"d9c5cc87-c132-43da-a319-1c65acf3c714","Type":"ContainerStarted","Data":"d6729f410731ff43faafb8bf0b010a43bfabfef95c20e86a0995e0d37f720308"} Oct 02 07:38:33 crc kubenswrapper[4829]: I1002 07:38:33.055081 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:33 crc kubenswrapper[4829]: I1002 07:38:33.101868 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" podStartSLOduration=4.1018412 podStartE2EDuration="4.1018412s" podCreationTimestamp="2025-10-02 07:38:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:38:33.094729664 +0000 UTC m=+1304.434378109" watchObservedRunningTime="2025-10-02 07:38:33.1018412 +0000 UTC m=+1304.441489635" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.221777 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.291084 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6wqhd"] Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.291394 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerName="dnsmasq-dns" containerID="cri-o://b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734" gracePeriod=10 Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.481600 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-cvpg4"] Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.484082 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.505582 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-cvpg4"] Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546037 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546100 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546429 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-config\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546516 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt7nv\" (UniqueName: \"kubernetes.io/projected/0c1b029b-bf2c-4b4b-b71c-6050f640212b-kube-api-access-qt7nv\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546658 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546730 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.546876 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650109 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-config\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650157 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt7nv\" (UniqueName: \"kubernetes.io/projected/0c1b029b-bf2c-4b4b-b71c-6050f640212b-kube-api-access-qt7nv\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650193 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650215 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650344 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650378 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.650414 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.651069 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-config\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.651366 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.651686 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.652649 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.652983 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.653546 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c1b029b-bf2c-4b4b-b71c-6050f640212b-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.676778 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt7nv\" (UniqueName: \"kubernetes.io/projected/0c1b029b-bf2c-4b4b-b71c-6050f640212b-kube-api-access-qt7nv\") pod \"dnsmasq-dns-6cd9bffc9-cvpg4\" (UID: \"0c1b029b-bf2c-4b4b-b71c-6050f640212b\") " pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.848950 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.856347 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.958292 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-config\") pod \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.958582 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-swift-storage-0\") pod \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.958611 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-svc\") pod \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.958644 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-sb\") pod \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.958686 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-nb\") pod \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.958728 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl2mz\" (UniqueName: \"kubernetes.io/projected/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-kube-api-access-jl2mz\") pod \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\" (UID: \"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c\") " Oct 02 07:38:40 crc kubenswrapper[4829]: I1002 07:38:40.963082 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-kube-api-access-jl2mz" (OuterVolumeSpecName: "kube-api-access-jl2mz") pod "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" (UID: "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c"). InnerVolumeSpecName "kube-api-access-jl2mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.020383 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" (UID: "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.041690 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" (UID: "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.041832 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" (UID: "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.042654 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-config" (OuterVolumeSpecName: "config") pod "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" (UID: "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.060964 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.061001 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.061014 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.061024 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.061037 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl2mz\" (UniqueName: \"kubernetes.io/projected/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-kube-api-access-jl2mz\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.080897 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" (UID: "c31152ee-d75a-4c89-89c4-a3b37bdbaa7c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.162847 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.169850 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.169871 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" event={"ID":"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c","Type":"ContainerDied","Data":"b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734"} Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.169945 4829 scope.go:117] "RemoveContainer" containerID="b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.169773 4829 generic.go:334] "Generic (PLEG): container finished" podID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerID="b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734" exitCode=0 Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.173356 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-6wqhd" event={"ID":"c31152ee-d75a-4c89-89c4-a3b37bdbaa7c","Type":"ContainerDied","Data":"1d91e88f78528cadd778e5e081befb332a44d930fe55bf5f60acb2e9f7cc3658"} Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.194587 4829 scope.go:117] "RemoveContainer" containerID="1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.219822 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6wqhd"] Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.227919 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-6wqhd"] Oct 02 07:38:41 crc kubenswrapper[4829]: E1002 07:38:41.233758 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc31152ee_d75a_4c89_89c4_a3b37bdbaa7c.slice/crio-1d91e88f78528cadd778e5e081befb332a44d930fe55bf5f60acb2e9f7cc3658\": RecentStats: unable to find data in memory cache]" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.260688 4829 scope.go:117] "RemoveContainer" containerID="b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734" Oct 02 07:38:41 crc kubenswrapper[4829]: E1002 07:38:41.261149 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734\": container with ID starting with b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734 not found: ID does not exist" containerID="b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.261196 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734"} err="failed to get container status \"b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734\": rpc error: code = NotFound desc = could not find container \"b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734\": container with ID starting with b76d4fd997aa2925da2ef596cc5dacf255a3ee1b8aace331992b5ce571a15734 not found: ID does not exist" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.261267 4829 scope.go:117] "RemoveContainer" containerID="1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204" Oct 02 07:38:41 crc kubenswrapper[4829]: E1002 07:38:41.261588 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204\": container with ID starting with 1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204 not found: ID does not exist" containerID="1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204" Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.261636 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204"} err="failed to get container status \"1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204\": rpc error: code = NotFound desc = could not find container \"1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204\": container with ID starting with 1a1c249045f8d27f6834024a4874847c1dc30bea136d4cc62192e2ff6958f204 not found: ID does not exist" Oct 02 07:38:41 crc kubenswrapper[4829]: W1002 07:38:41.317843 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c1b029b_bf2c_4b4b_b71c_6050f640212b.slice/crio-4c2c7a7d48da1dfb86aac743df2e90ef69b29f026c107ae7544909bb3adc364f WatchSource:0}: Error finding container 4c2c7a7d48da1dfb86aac743df2e90ef69b29f026c107ae7544909bb3adc364f: Status 404 returned error can't find the container with id 4c2c7a7d48da1dfb86aac743df2e90ef69b29f026c107ae7544909bb3adc364f Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.319729 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-cvpg4"] Oct 02 07:38:41 crc kubenswrapper[4829]: I1002 07:38:41.477156 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" path="/var/lib/kubelet/pods/c31152ee-d75a-4c89-89c4-a3b37bdbaa7c/volumes" Oct 02 07:38:42 crc kubenswrapper[4829]: I1002 07:38:42.183612 4829 generic.go:334] "Generic (PLEG): container finished" podID="0c1b029b-bf2c-4b4b-b71c-6050f640212b" containerID="17c580d9d01c064433a7575319485bec3fe6b5eb1b8dbcd4f23a95642543fa68" exitCode=0 Oct 02 07:38:42 crc kubenswrapper[4829]: I1002 07:38:42.183659 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" event={"ID":"0c1b029b-bf2c-4b4b-b71c-6050f640212b","Type":"ContainerDied","Data":"17c580d9d01c064433a7575319485bec3fe6b5eb1b8dbcd4f23a95642543fa68"} Oct 02 07:38:42 crc kubenswrapper[4829]: I1002 07:38:42.183684 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" event={"ID":"0c1b029b-bf2c-4b4b-b71c-6050f640212b","Type":"ContainerStarted","Data":"4c2c7a7d48da1dfb86aac743df2e90ef69b29f026c107ae7544909bb3adc364f"} Oct 02 07:38:43 crc kubenswrapper[4829]: I1002 07:38:43.195749 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" event={"ID":"0c1b029b-bf2c-4b4b-b71c-6050f640212b","Type":"ContainerStarted","Data":"00191d0c2ad279dc6d53e25ddd955362af165c6d92c7dec2cf116303280e2a7e"} Oct 02 07:38:43 crc kubenswrapper[4829]: I1002 07:38:43.196011 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:43 crc kubenswrapper[4829]: I1002 07:38:43.226082 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" podStartSLOduration=3.226027748 podStartE2EDuration="3.226027748s" podCreationTimestamp="2025-10-02 07:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:38:43.213697856 +0000 UTC m=+1314.553346281" watchObservedRunningTime="2025-10-02 07:38:43.226027748 +0000 UTC m=+1314.565676153" Oct 02 07:38:50 crc kubenswrapper[4829]: I1002 07:38:50.851504 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cd9bffc9-cvpg4" Oct 02 07:38:50 crc kubenswrapper[4829]: I1002 07:38:50.933700 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-c7fhz"] Oct 02 07:38:50 crc kubenswrapper[4829]: I1002 07:38:50.934044 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerName="dnsmasq-dns" containerID="cri-o://d6729f410731ff43faafb8bf0b010a43bfabfef95c20e86a0995e0d37f720308" gracePeriod=10 Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.284097 4829 generic.go:334] "Generic (PLEG): container finished" podID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerID="d6729f410731ff43faafb8bf0b010a43bfabfef95c20e86a0995e0d37f720308" exitCode=0 Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.284493 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" event={"ID":"d9c5cc87-c132-43da-a319-1c65acf3c714","Type":"ContainerDied","Data":"d6729f410731ff43faafb8bf0b010a43bfabfef95c20e86a0995e0d37f720308"} Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.472782 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.605727 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-nb\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.605798 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2vhl\" (UniqueName: \"kubernetes.io/projected/d9c5cc87-c132-43da-a319-1c65acf3c714-kube-api-access-k2vhl\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.605947 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-svc\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.605974 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-swift-storage-0\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.605995 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-openstack-edpm-ipam\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.606039 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-sb\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.606143 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-config\") pod \"d9c5cc87-c132-43da-a319-1c65acf3c714\" (UID: \"d9c5cc87-c132-43da-a319-1c65acf3c714\") " Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.656738 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c5cc87-c132-43da-a319-1c65acf3c714-kube-api-access-k2vhl" (OuterVolumeSpecName: "kube-api-access-k2vhl") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "kube-api-access-k2vhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.708762 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2vhl\" (UniqueName: \"kubernetes.io/projected/d9c5cc87-c132-43da-a319-1c65acf3c714-kube-api-access-k2vhl\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.734301 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-config" (OuterVolumeSpecName: "config") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.735402 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.782100 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.782806 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.790259 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.791701 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d9c5cc87-c132-43da-a319-1c65acf3c714" (UID: "d9c5cc87-c132-43da-a319-1c65acf3c714"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.810073 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.810112 4829 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.810125 4829 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.810137 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.810148 4829 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:51 crc kubenswrapper[4829]: I1002 07:38:51.810158 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9c5cc87-c132-43da-a319-1c65acf3c714-config\") on node \"crc\" DevicePath \"\"" Oct 02 07:38:52 crc kubenswrapper[4829]: I1002 07:38:52.295423 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" event={"ID":"d9c5cc87-c132-43da-a319-1c65acf3c714","Type":"ContainerDied","Data":"973fb12556b7ab018a2170eacc91f13e7a13f121cce77bd0ba42fed58b66a80c"} Oct 02 07:38:52 crc kubenswrapper[4829]: I1002 07:38:52.295478 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-c7fhz" Oct 02 07:38:52 crc kubenswrapper[4829]: I1002 07:38:52.295525 4829 scope.go:117] "RemoveContainer" containerID="d6729f410731ff43faafb8bf0b010a43bfabfef95c20e86a0995e0d37f720308" Oct 02 07:38:52 crc kubenswrapper[4829]: I1002 07:38:52.328910 4829 scope.go:117] "RemoveContainer" containerID="a17d896a685a3e3ba84ea3fb874c02c80d8b3106858f148c4322939e9143e154" Oct 02 07:38:52 crc kubenswrapper[4829]: I1002 07:38:52.338173 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-c7fhz"] Oct 02 07:38:52 crc kubenswrapper[4829]: I1002 07:38:52.358671 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-c7fhz"] Oct 02 07:38:53 crc kubenswrapper[4829]: I1002 07:38:53.474511 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" path="/var/lib/kubelet/pods/d9c5cc87-c132-43da-a319-1c65acf3c714/volumes" Oct 02 07:38:55 crc kubenswrapper[4829]: I1002 07:38:55.329313 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:38:55 crc kubenswrapper[4829]: I1002 07:38:55.329589 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:38:55 crc kubenswrapper[4829]: I1002 07:38:55.329639 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:38:55 crc kubenswrapper[4829]: I1002 07:38:55.330362 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17a95fceb29822d3307537c1facf40f10ec266fc571083711123b45f3e6f1880"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:38:55 crc kubenswrapper[4829]: I1002 07:38:55.330410 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://17a95fceb29822d3307537c1facf40f10ec266fc571083711123b45f3e6f1880" gracePeriod=600 Oct 02 07:38:56 crc kubenswrapper[4829]: I1002 07:38:56.341357 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="17a95fceb29822d3307537c1facf40f10ec266fc571083711123b45f3e6f1880" exitCode=0 Oct 02 07:38:56 crc kubenswrapper[4829]: I1002 07:38:56.341381 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"17a95fceb29822d3307537c1facf40f10ec266fc571083711123b45f3e6f1880"} Oct 02 07:38:56 crc kubenswrapper[4829]: I1002 07:38:56.341894 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31"} Oct 02 07:38:56 crc kubenswrapper[4829]: I1002 07:38:56.341910 4829 scope.go:117] "RemoveContainer" containerID="ae2b08236229e74b63b3e0438daf20903c2e9ff5c5e97936f06ade85ab34a7a5" Oct 02 07:39:02 crc kubenswrapper[4829]: I1002 07:39:02.422434 4829 generic.go:334] "Generic (PLEG): container finished" podID="b66249bb-0358-455a-8ca1-7c9688cfe5f0" containerID="d168b697068a6173b56c693ccb7c2971c88697528998bc48a515ed1e792e3142" exitCode=0 Oct 02 07:39:02 crc kubenswrapper[4829]: I1002 07:39:02.422539 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66249bb-0358-455a-8ca1-7c9688cfe5f0","Type":"ContainerDied","Data":"d168b697068a6173b56c693ccb7c2971c88697528998bc48a515ed1e792e3142"} Oct 02 07:39:02 crc kubenswrapper[4829]: I1002 07:39:02.425761 4829 generic.go:334] "Generic (PLEG): container finished" podID="43522a1c-881d-4d33-91e3-3dea8c241d8c" containerID="5fb7cef4886633b52a13a40754ce132a0375ed51da9eb62e32d41135c0bd1d0f" exitCode=0 Oct 02 07:39:02 crc kubenswrapper[4829]: I1002 07:39:02.425799 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43522a1c-881d-4d33-91e3-3dea8c241d8c","Type":"ContainerDied","Data":"5fb7cef4886633b52a13a40754ce132a0375ed51da9eb62e32d41135c0bd1d0f"} Oct 02 07:39:03 crc kubenswrapper[4829]: I1002 07:39:03.436240 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"43522a1c-881d-4d33-91e3-3dea8c241d8c","Type":"ContainerStarted","Data":"b5c9be8caacfb16ece88cbcbce5a1ddbeaf80b99aa27123a4543247c50c1dd4e"} Oct 02 07:39:03 crc kubenswrapper[4829]: I1002 07:39:03.436840 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 07:39:03 crc kubenswrapper[4829]: I1002 07:39:03.438677 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b66249bb-0358-455a-8ca1-7c9688cfe5f0","Type":"ContainerStarted","Data":"10ce4ed77e7bafef5acfbbf54cdea029de86bf05f6f7df0bb69c377f976a56e0"} Oct 02 07:39:03 crc kubenswrapper[4829]: I1002 07:39:03.438928 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:39:03 crc kubenswrapper[4829]: I1002 07:39:03.483129 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.48311111 podStartE2EDuration="39.48311111s" podCreationTimestamp="2025-10-02 07:38:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:39:03.454031818 +0000 UTC m=+1334.793680223" watchObservedRunningTime="2025-10-02 07:39:03.48311111 +0000 UTC m=+1334.822759515" Oct 02 07:39:03 crc kubenswrapper[4829]: I1002 07:39:03.504468 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.504449726 podStartE2EDuration="37.504449726s" podCreationTimestamp="2025-10-02 07:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 07:39:03.501007637 +0000 UTC m=+1334.840656052" watchObservedRunningTime="2025-10-02 07:39:03.504449726 +0000 UTC m=+1334.844098131" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.992218 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26"] Oct 02 07:39:04 crc kubenswrapper[4829]: E1002 07:39:04.995373 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerName="init" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.995665 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerName="init" Oct 02 07:39:04 crc kubenswrapper[4829]: E1002 07:39:04.995701 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerName="init" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.995708 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerName="init" Oct 02 07:39:04 crc kubenswrapper[4829]: E1002 07:39:04.995715 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerName="dnsmasq-dns" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.995721 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerName="dnsmasq-dns" Oct 02 07:39:04 crc kubenswrapper[4829]: E1002 07:39:04.995733 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerName="dnsmasq-dns" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.995738 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerName="dnsmasq-dns" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.996064 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c5cc87-c132-43da-a319-1c65acf3c714" containerName="dnsmasq-dns" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.996089 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31152ee-d75a-4c89-89c4-a3b37bdbaa7c" containerName="dnsmasq-dns" Oct 02 07:39:04 crc kubenswrapper[4829]: I1002 07:39:04.996884 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.003622 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.003991 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.004864 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.010445 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.023043 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26"] Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.096353 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.096489 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.096540 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g89vw\" (UniqueName: \"kubernetes.io/projected/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-kube-api-access-g89vw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.096740 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.198519 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g89vw\" (UniqueName: \"kubernetes.io/projected/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-kube-api-access-g89vw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.198721 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.198870 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.199112 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.205778 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.206043 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.206791 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.216182 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g89vw\" (UniqueName: \"kubernetes.io/projected/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-kube-api-access-g89vw\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.334746 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:05 crc kubenswrapper[4829]: I1002 07:39:05.861809 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26"] Oct 02 07:39:06 crc kubenswrapper[4829]: I1002 07:39:06.472528 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" event={"ID":"01ec7e0b-c62e-4bd4-ab7e-6756c5314161","Type":"ContainerStarted","Data":"fc1059cb111fd6dcb086cadb049e2d368db83738516571617f7865d298edef33"} Oct 02 07:39:15 crc kubenswrapper[4829]: I1002 07:39:15.579732 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 07:39:15 crc kubenswrapper[4829]: I1002 07:39:15.588524 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" event={"ID":"01ec7e0b-c62e-4bd4-ab7e-6756c5314161","Type":"ContainerStarted","Data":"16ac161afd7d7d528513499c5ce10c8de5fe4ea43cbb5e55d26e2a3c2f2e5b26"} Oct 02 07:39:15 crc kubenswrapper[4829]: I1002 07:39:15.631296 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" podStartSLOduration=2.957545089 podStartE2EDuration="11.631280364s" podCreationTimestamp="2025-10-02 07:39:04 +0000 UTC" firstStartedPulling="2025-10-02 07:39:05.881347251 +0000 UTC m=+1337.220995656" lastFinishedPulling="2025-10-02 07:39:14.555082496 +0000 UTC m=+1345.894730931" observedRunningTime="2025-10-02 07:39:15.630949864 +0000 UTC m=+1346.970598269" watchObservedRunningTime="2025-10-02 07:39:15.631280364 +0000 UTC m=+1346.970928769" Oct 02 07:39:17 crc kubenswrapper[4829]: I1002 07:39:17.661350 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 07:39:26 crc kubenswrapper[4829]: I1002 07:39:26.704433 4829 generic.go:334] "Generic (PLEG): container finished" podID="01ec7e0b-c62e-4bd4-ab7e-6756c5314161" containerID="16ac161afd7d7d528513499c5ce10c8de5fe4ea43cbb5e55d26e2a3c2f2e5b26" exitCode=0 Oct 02 07:39:26 crc kubenswrapper[4829]: I1002 07:39:26.704531 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" event={"ID":"01ec7e0b-c62e-4bd4-ab7e-6756c5314161","Type":"ContainerDied","Data":"16ac161afd7d7d528513499c5ce10c8de5fe4ea43cbb5e55d26e2a3c2f2e5b26"} Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.320429 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.323900 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g89vw\" (UniqueName: \"kubernetes.io/projected/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-kube-api-access-g89vw\") pod \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.324038 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-ssh-key\") pod \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.324065 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-repo-setup-combined-ca-bundle\") pod \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.324109 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-inventory\") pod \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\" (UID: \"01ec7e0b-c62e-4bd4-ab7e-6756c5314161\") " Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.330893 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-kube-api-access-g89vw" (OuterVolumeSpecName: "kube-api-access-g89vw") pod "01ec7e0b-c62e-4bd4-ab7e-6756c5314161" (UID: "01ec7e0b-c62e-4bd4-ab7e-6756c5314161"). InnerVolumeSpecName "kube-api-access-g89vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.330912 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "01ec7e0b-c62e-4bd4-ab7e-6756c5314161" (UID: "01ec7e0b-c62e-4bd4-ab7e-6756c5314161"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.392121 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "01ec7e0b-c62e-4bd4-ab7e-6756c5314161" (UID: "01ec7e0b-c62e-4bd4-ab7e-6756c5314161"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.396505 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-inventory" (OuterVolumeSpecName: "inventory") pod "01ec7e0b-c62e-4bd4-ab7e-6756c5314161" (UID: "01ec7e0b-c62e-4bd4-ab7e-6756c5314161"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.430108 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.430147 4829 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.430165 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.430179 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g89vw\" (UniqueName: \"kubernetes.io/projected/01ec7e0b-c62e-4bd4-ab7e-6756c5314161-kube-api-access-g89vw\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.733206 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" event={"ID":"01ec7e0b-c62e-4bd4-ab7e-6756c5314161","Type":"ContainerDied","Data":"fc1059cb111fd6dcb086cadb049e2d368db83738516571617f7865d298edef33"} Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.733300 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc1059cb111fd6dcb086cadb049e2d368db83738516571617f7865d298edef33" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.733254 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.834889 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777"] Oct 02 07:39:28 crc kubenswrapper[4829]: E1002 07:39:28.835774 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01ec7e0b-c62e-4bd4-ab7e-6756c5314161" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.835812 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="01ec7e0b-c62e-4bd4-ab7e-6756c5314161" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.836167 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="01ec7e0b-c62e-4bd4-ab7e-6756c5314161" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.837284 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.839367 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.839678 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.840249 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.840388 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.864768 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777"] Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.939011 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.939293 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:28 crc kubenswrapper[4829]: I1002 07:39:28.939667 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j746\" (UniqueName: \"kubernetes.io/projected/fb03e152-ef79-4d7f-920d-5d1fbacb7167-kube-api-access-2j746\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.042854 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j746\" (UniqueName: \"kubernetes.io/projected/fb03e152-ef79-4d7f-920d-5d1fbacb7167-kube-api-access-2j746\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.043434 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.043650 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.047132 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.048622 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.071998 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j746\" (UniqueName: \"kubernetes.io/projected/fb03e152-ef79-4d7f-920d-5d1fbacb7167-kube-api-access-2j746\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fm777\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.158479 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:29 crc kubenswrapper[4829]: I1002 07:39:29.761195 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777"] Oct 02 07:39:30 crc kubenswrapper[4829]: I1002 07:39:30.757204 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" event={"ID":"fb03e152-ef79-4d7f-920d-5d1fbacb7167","Type":"ContainerStarted","Data":"e8939194260caa135a78cdd5e943112b3847843a330cfd8b976f45e83906c8ab"} Oct 02 07:39:30 crc kubenswrapper[4829]: I1002 07:39:30.757628 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" event={"ID":"fb03e152-ef79-4d7f-920d-5d1fbacb7167","Type":"ContainerStarted","Data":"7d27a0aa908d20b7a36d0091c11cfae13b6fd8231a2f5e210518b8fecc70238b"} Oct 02 07:39:30 crc kubenswrapper[4829]: I1002 07:39:30.784329 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" podStartSLOduration=2.372253449 podStartE2EDuration="2.784308412s" podCreationTimestamp="2025-10-02 07:39:28 +0000 UTC" firstStartedPulling="2025-10-02 07:39:29.764476701 +0000 UTC m=+1361.104125106" lastFinishedPulling="2025-10-02 07:39:30.176531624 +0000 UTC m=+1361.516180069" observedRunningTime="2025-10-02 07:39:30.777504767 +0000 UTC m=+1362.117153172" watchObservedRunningTime="2025-10-02 07:39:30.784308412 +0000 UTC m=+1362.123956837" Oct 02 07:39:33 crc kubenswrapper[4829]: I1002 07:39:33.784791 4829 generic.go:334] "Generic (PLEG): container finished" podID="fb03e152-ef79-4d7f-920d-5d1fbacb7167" containerID="e8939194260caa135a78cdd5e943112b3847843a330cfd8b976f45e83906c8ab" exitCode=0 Oct 02 07:39:33 crc kubenswrapper[4829]: I1002 07:39:33.784910 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" event={"ID":"fb03e152-ef79-4d7f-920d-5d1fbacb7167","Type":"ContainerDied","Data":"e8939194260caa135a78cdd5e943112b3847843a330cfd8b976f45e83906c8ab"} Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.289746 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.387203 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-inventory\") pod \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.387377 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-ssh-key\") pod \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.387511 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j746\" (UniqueName: \"kubernetes.io/projected/fb03e152-ef79-4d7f-920d-5d1fbacb7167-kube-api-access-2j746\") pod \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\" (UID: \"fb03e152-ef79-4d7f-920d-5d1fbacb7167\") " Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.395525 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb03e152-ef79-4d7f-920d-5d1fbacb7167-kube-api-access-2j746" (OuterVolumeSpecName: "kube-api-access-2j746") pod "fb03e152-ef79-4d7f-920d-5d1fbacb7167" (UID: "fb03e152-ef79-4d7f-920d-5d1fbacb7167"). InnerVolumeSpecName "kube-api-access-2j746". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.415496 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-inventory" (OuterVolumeSpecName: "inventory") pod "fb03e152-ef79-4d7f-920d-5d1fbacb7167" (UID: "fb03e152-ef79-4d7f-920d-5d1fbacb7167"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.418137 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fb03e152-ef79-4d7f-920d-5d1fbacb7167" (UID: "fb03e152-ef79-4d7f-920d-5d1fbacb7167"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.490317 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.490368 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j746\" (UniqueName: \"kubernetes.io/projected/fb03e152-ef79-4d7f-920d-5d1fbacb7167-kube-api-access-2j746\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.490393 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fb03e152-ef79-4d7f-920d-5d1fbacb7167-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.808182 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" event={"ID":"fb03e152-ef79-4d7f-920d-5d1fbacb7167","Type":"ContainerDied","Data":"7d27a0aa908d20b7a36d0091c11cfae13b6fd8231a2f5e210518b8fecc70238b"} Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.808270 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fm777" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.808287 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d27a0aa908d20b7a36d0091c11cfae13b6fd8231a2f5e210518b8fecc70238b" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.944406 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk"] Oct 02 07:39:35 crc kubenswrapper[4829]: E1002 07:39:35.944915 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb03e152-ef79-4d7f-920d-5d1fbacb7167" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.944940 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb03e152-ef79-4d7f-920d-5d1fbacb7167" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.945142 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb03e152-ef79-4d7f-920d-5d1fbacb7167" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.945950 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.948394 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.948648 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.948675 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.948863 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:39:35 crc kubenswrapper[4829]: I1002 07:39:35.958034 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk"] Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.030674 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.031203 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.031344 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.031467 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6jsw\" (UniqueName: \"kubernetes.io/projected/46cccbf0-9925-4708-bd05-119652ca2732-kube-api-access-g6jsw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.133959 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.134310 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.134421 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6jsw\" (UniqueName: \"kubernetes.io/projected/46cccbf0-9925-4708-bd05-119652ca2732-kube-api-access-g6jsw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.134589 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.143952 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.157482 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.158610 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.159686 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6jsw\" (UniqueName: \"kubernetes.io/projected/46cccbf0-9925-4708-bd05-119652ca2732-kube-api-access-g6jsw\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.272669 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.780464 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk"] Oct 02 07:39:36 crc kubenswrapper[4829]: W1002 07:39:36.781006 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46cccbf0_9925_4708_bd05_119652ca2732.slice/crio-c7a679d3bf8232b49518ad312f1ee5a417457987775a57d254111f05020009a8 WatchSource:0}: Error finding container c7a679d3bf8232b49518ad312f1ee5a417457987775a57d254111f05020009a8: Status 404 returned error can't find the container with id c7a679d3bf8232b49518ad312f1ee5a417457987775a57d254111f05020009a8 Oct 02 07:39:36 crc kubenswrapper[4829]: I1002 07:39:36.821738 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" event={"ID":"46cccbf0-9925-4708-bd05-119652ca2732","Type":"ContainerStarted","Data":"c7a679d3bf8232b49518ad312f1ee5a417457987775a57d254111f05020009a8"} Oct 02 07:39:37 crc kubenswrapper[4829]: I1002 07:39:37.836439 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" event={"ID":"46cccbf0-9925-4708-bd05-119652ca2732","Type":"ContainerStarted","Data":"b40e94198255c6162b1d7187f56abc0eb1399e84c0a663316325dcef2e7a0dea"} Oct 02 07:39:37 crc kubenswrapper[4829]: I1002 07:39:37.868152 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" podStartSLOduration=2.264447123 podStartE2EDuration="2.868131622s" podCreationTimestamp="2025-10-02 07:39:35 +0000 UTC" firstStartedPulling="2025-10-02 07:39:36.784688833 +0000 UTC m=+1368.124337238" lastFinishedPulling="2025-10-02 07:39:37.388373292 +0000 UTC m=+1368.728021737" observedRunningTime="2025-10-02 07:39:37.862634458 +0000 UTC m=+1369.202282883" watchObservedRunningTime="2025-10-02 07:39:37.868131622 +0000 UTC m=+1369.207780037" Oct 02 07:39:56 crc kubenswrapper[4829]: I1002 07:39:56.025185 4829 scope.go:117] "RemoveContainer" containerID="9c33059abbde95491ae84e5c6e8fb89d301e4b1d50816cd2a36e60454d72a0c8" Oct 02 07:39:56 crc kubenswrapper[4829]: I1002 07:39:56.060496 4829 scope.go:117] "RemoveContainer" containerID="a8d3d630c1db387878ae8bb1b1baed1a5515ac74b05ca5b0f1313dc3793d0171" Oct 02 07:39:56 crc kubenswrapper[4829]: I1002 07:39:56.088185 4829 scope.go:117] "RemoveContainer" containerID="4fd0cc46b9bf567752d33e830edff9203b00336e0a8ea83b189d5c929c40395a" Oct 02 07:39:56 crc kubenswrapper[4829]: I1002 07:39:56.127841 4829 scope.go:117] "RemoveContainer" containerID="91bbbebef21bd98ba4714faafc6d34cf2ceca44eff5e5ec17afa2d02308bef43" Oct 02 07:39:56 crc kubenswrapper[4829]: I1002 07:39:56.158721 4829 scope.go:117] "RemoveContainer" containerID="bcd7a8bf4486d27e6f4d8427b9c5722ebfeb26a8149952713aa27fcdb806e982" Oct 02 07:40:53 crc kubenswrapper[4829]: I1002 07:40:53.704460 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fnx7g"] Oct 02 07:40:53 crc kubenswrapper[4829]: I1002 07:40:53.710446 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:53 crc kubenswrapper[4829]: I1002 07:40:53.738071 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fnx7g"] Oct 02 07:40:53 crc kubenswrapper[4829]: I1002 07:40:53.900899 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvf8b\" (UniqueName: \"kubernetes.io/projected/08ff0b4a-f70b-4cba-9791-74dd0182aa54-kube-api-access-kvf8b\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:53 crc kubenswrapper[4829]: I1002 07:40:53.901040 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-utilities\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:53 crc kubenswrapper[4829]: I1002 07:40:53.901315 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-catalog-content\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.003160 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-catalog-content\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.003332 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvf8b\" (UniqueName: \"kubernetes.io/projected/08ff0b4a-f70b-4cba-9791-74dd0182aa54-kube-api-access-kvf8b\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.003398 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-utilities\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.003786 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-catalog-content\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.003851 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-utilities\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.029852 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvf8b\" (UniqueName: \"kubernetes.io/projected/08ff0b4a-f70b-4cba-9791-74dd0182aa54-kube-api-access-kvf8b\") pod \"redhat-operators-fnx7g\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.039370 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.540420 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fnx7g"] Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.762366 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerStarted","Data":"0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066"} Oct 02 07:40:54 crc kubenswrapper[4829]: I1002 07:40:54.762697 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerStarted","Data":"63861e8e9b3011a235be237d83146a3a77ad5932cbb3126e64bdb3e1e0e9c5ad"} Oct 02 07:40:55 crc kubenswrapper[4829]: I1002 07:40:55.328965 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:40:55 crc kubenswrapper[4829]: I1002 07:40:55.329058 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:40:55 crc kubenswrapper[4829]: I1002 07:40:55.787575 4829 generic.go:334] "Generic (PLEG): container finished" podID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerID="0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066" exitCode=0 Oct 02 07:40:55 crc kubenswrapper[4829]: I1002 07:40:55.787851 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerDied","Data":"0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066"} Oct 02 07:40:56 crc kubenswrapper[4829]: I1002 07:40:56.254213 4829 scope.go:117] "RemoveContainer" containerID="dae3209bf1f7bb8956a52297b8d81605ad2f3b0ca6b90bc870590478cc3ac44f" Oct 02 07:40:56 crc kubenswrapper[4829]: I1002 07:40:56.468998 4829 scope.go:117] "RemoveContainer" containerID="229bc912bc4f9bd9d62366c21ed8e572caa493456101b39333c2c03734cdb5d9" Oct 02 07:40:56 crc kubenswrapper[4829]: I1002 07:40:56.522995 4829 scope.go:117] "RemoveContainer" containerID="9814c7fd97f5c331f361eb5c9bf44206b3cba730dc3c445f6f3835986ffb7bef" Oct 02 07:40:56 crc kubenswrapper[4829]: I1002 07:40:56.802768 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerStarted","Data":"86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac"} Oct 02 07:40:57 crc kubenswrapper[4829]: I1002 07:40:57.819180 4829 generic.go:334] "Generic (PLEG): container finished" podID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerID="86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac" exitCode=0 Oct 02 07:40:57 crc kubenswrapper[4829]: I1002 07:40:57.819251 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerDied","Data":"86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac"} Oct 02 07:40:58 crc kubenswrapper[4829]: I1002 07:40:58.834724 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerStarted","Data":"5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f"} Oct 02 07:40:58 crc kubenswrapper[4829]: I1002 07:40:58.856422 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fnx7g" podStartSLOduration=3.310939581 podStartE2EDuration="5.856406796s" podCreationTimestamp="2025-10-02 07:40:53 +0000 UTC" firstStartedPulling="2025-10-02 07:40:55.789720027 +0000 UTC m=+1447.129368432" lastFinishedPulling="2025-10-02 07:40:58.335187242 +0000 UTC m=+1449.674835647" observedRunningTime="2025-10-02 07:40:58.853756151 +0000 UTC m=+1450.193404566" watchObservedRunningTime="2025-10-02 07:40:58.856406796 +0000 UTC m=+1450.196055191" Oct 02 07:41:04 crc kubenswrapper[4829]: I1002 07:41:04.046074 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:41:04 crc kubenswrapper[4829]: I1002 07:41:04.046823 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:41:05 crc kubenswrapper[4829]: I1002 07:41:05.109535 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fnx7g" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="registry-server" probeResult="failure" output=< Oct 02 07:41:05 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:41:05 crc kubenswrapper[4829]: > Oct 02 07:41:14 crc kubenswrapper[4829]: I1002 07:41:14.104842 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:41:14 crc kubenswrapper[4829]: I1002 07:41:14.160071 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:41:14 crc kubenswrapper[4829]: I1002 07:41:14.337408 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fnx7g"] Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.019385 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fnx7g" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="registry-server" containerID="cri-o://5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f" gracePeriod=2 Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.564635 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.603961 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-utilities\") pod \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.604067 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-catalog-content\") pod \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.604297 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvf8b\" (UniqueName: \"kubernetes.io/projected/08ff0b4a-f70b-4cba-9791-74dd0182aa54-kube-api-access-kvf8b\") pod \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\" (UID: \"08ff0b4a-f70b-4cba-9791-74dd0182aa54\") " Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.605331 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-utilities" (OuterVolumeSpecName: "utilities") pod "08ff0b4a-f70b-4cba-9791-74dd0182aa54" (UID: "08ff0b4a-f70b-4cba-9791-74dd0182aa54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.616013 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ff0b4a-f70b-4cba-9791-74dd0182aa54-kube-api-access-kvf8b" (OuterVolumeSpecName: "kube-api-access-kvf8b") pod "08ff0b4a-f70b-4cba-9791-74dd0182aa54" (UID: "08ff0b4a-f70b-4cba-9791-74dd0182aa54"). InnerVolumeSpecName "kube-api-access-kvf8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.692370 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08ff0b4a-f70b-4cba-9791-74dd0182aa54" (UID: "08ff0b4a-f70b-4cba-9791-74dd0182aa54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.707978 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.708020 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08ff0b4a-f70b-4cba-9791-74dd0182aa54-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:16 crc kubenswrapper[4829]: I1002 07:41:16.708033 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvf8b\" (UniqueName: \"kubernetes.io/projected/08ff0b4a-f70b-4cba-9791-74dd0182aa54-kube-api-access-kvf8b\") on node \"crc\" DevicePath \"\"" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.029746 4829 generic.go:334] "Generic (PLEG): container finished" podID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerID="5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f" exitCode=0 Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.029785 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerDied","Data":"5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f"} Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.029817 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fnx7g" event={"ID":"08ff0b4a-f70b-4cba-9791-74dd0182aa54","Type":"ContainerDied","Data":"63861e8e9b3011a235be237d83146a3a77ad5932cbb3126e64bdb3e1e0e9c5ad"} Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.029821 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fnx7g" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.029835 4829 scope.go:117] "RemoveContainer" containerID="5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.057167 4829 scope.go:117] "RemoveContainer" containerID="86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.068672 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fnx7g"] Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.079648 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fnx7g"] Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.095828 4829 scope.go:117] "RemoveContainer" containerID="0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.156440 4829 scope.go:117] "RemoveContainer" containerID="5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f" Oct 02 07:41:17 crc kubenswrapper[4829]: E1002 07:41:17.157154 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f\": container with ID starting with 5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f not found: ID does not exist" containerID="5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.157198 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f"} err="failed to get container status \"5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f\": rpc error: code = NotFound desc = could not find container \"5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f\": container with ID starting with 5141e5edf0a2aaf963f4f4fdc873b7d6b41d8f06710955ebcf88eb9bd653c06f not found: ID does not exist" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.157244 4829 scope.go:117] "RemoveContainer" containerID="86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac" Oct 02 07:41:17 crc kubenswrapper[4829]: E1002 07:41:17.159131 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac\": container with ID starting with 86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac not found: ID does not exist" containerID="86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.159175 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac"} err="failed to get container status \"86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac\": rpc error: code = NotFound desc = could not find container \"86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac\": container with ID starting with 86682ee96fb498eb3a3df5240819b597ad513f628e490a4b3c2278e7ac8a4eac not found: ID does not exist" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.159202 4829 scope.go:117] "RemoveContainer" containerID="0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066" Oct 02 07:41:17 crc kubenswrapper[4829]: E1002 07:41:17.159790 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066\": container with ID starting with 0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066 not found: ID does not exist" containerID="0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.159838 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066"} err="failed to get container status \"0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066\": rpc error: code = NotFound desc = could not find container \"0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066\": container with ID starting with 0373e313c97264ec0bbcc48b94605dc071b20edd3fb2db14bc1fbfa784f04066 not found: ID does not exist" Oct 02 07:41:17 crc kubenswrapper[4829]: I1002 07:41:17.472299 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" path="/var/lib/kubelet/pods/08ff0b4a-f70b-4cba-9791-74dd0182aa54/volumes" Oct 02 07:41:25 crc kubenswrapper[4829]: I1002 07:41:25.329729 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:41:25 crc kubenswrapper[4829]: I1002 07:41:25.330546 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:41:55 crc kubenswrapper[4829]: I1002 07:41:55.329688 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:41:55 crc kubenswrapper[4829]: I1002 07:41:55.330277 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:41:55 crc kubenswrapper[4829]: I1002 07:41:55.330333 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:41:55 crc kubenswrapper[4829]: I1002 07:41:55.331147 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:41:55 crc kubenswrapper[4829]: I1002 07:41:55.331216 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" gracePeriod=600 Oct 02 07:41:55 crc kubenswrapper[4829]: E1002 07:41:55.498517 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:41:56 crc kubenswrapper[4829]: I1002 07:41:56.481946 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" exitCode=0 Oct 02 07:41:56 crc kubenswrapper[4829]: I1002 07:41:56.482022 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31"} Oct 02 07:41:56 crc kubenswrapper[4829]: I1002 07:41:56.482089 4829 scope.go:117] "RemoveContainer" containerID="17a95fceb29822d3307537c1facf40f10ec266fc571083711123b45f3e6f1880" Oct 02 07:41:56 crc kubenswrapper[4829]: I1002 07:41:56.483102 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:41:56 crc kubenswrapper[4829]: E1002 07:41:56.483739 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:42:07 crc kubenswrapper[4829]: I1002 07:42:07.461641 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:42:07 crc kubenswrapper[4829]: E1002 07:42:07.462717 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:42:22 crc kubenswrapper[4829]: I1002 07:42:22.461176 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:42:22 crc kubenswrapper[4829]: E1002 07:42:22.461881 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:42:34 crc kubenswrapper[4829]: I1002 07:42:34.462528 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:42:34 crc kubenswrapper[4829]: E1002 07:42:34.463907 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:42:48 crc kubenswrapper[4829]: I1002 07:42:48.461144 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:42:48 crc kubenswrapper[4829]: E1002 07:42:48.461963 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:42:53 crc kubenswrapper[4829]: I1002 07:42:53.156719 4829 generic.go:334] "Generic (PLEG): container finished" podID="46cccbf0-9925-4708-bd05-119652ca2732" containerID="b40e94198255c6162b1d7187f56abc0eb1399e84c0a663316325dcef2e7a0dea" exitCode=0 Oct 02 07:42:53 crc kubenswrapper[4829]: I1002 07:42:53.156840 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" event={"ID":"46cccbf0-9925-4708-bd05-119652ca2732","Type":"ContainerDied","Data":"b40e94198255c6162b1d7187f56abc0eb1399e84c0a663316325dcef2e7a0dea"} Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.837551 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.946831 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-inventory\") pod \"46cccbf0-9925-4708-bd05-119652ca2732\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.946884 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-ssh-key\") pod \"46cccbf0-9925-4708-bd05-119652ca2732\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.946944 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-bootstrap-combined-ca-bundle\") pod \"46cccbf0-9925-4708-bd05-119652ca2732\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.946991 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6jsw\" (UniqueName: \"kubernetes.io/projected/46cccbf0-9925-4708-bd05-119652ca2732-kube-api-access-g6jsw\") pod \"46cccbf0-9925-4708-bd05-119652ca2732\" (UID: \"46cccbf0-9925-4708-bd05-119652ca2732\") " Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.953332 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "46cccbf0-9925-4708-bd05-119652ca2732" (UID: "46cccbf0-9925-4708-bd05-119652ca2732"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.959418 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46cccbf0-9925-4708-bd05-119652ca2732-kube-api-access-g6jsw" (OuterVolumeSpecName: "kube-api-access-g6jsw") pod "46cccbf0-9925-4708-bd05-119652ca2732" (UID: "46cccbf0-9925-4708-bd05-119652ca2732"). InnerVolumeSpecName "kube-api-access-g6jsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.980042 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "46cccbf0-9925-4708-bd05-119652ca2732" (UID: "46cccbf0-9925-4708-bd05-119652ca2732"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:54 crc kubenswrapper[4829]: I1002 07:42:54.985435 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-inventory" (OuterVolumeSpecName: "inventory") pod "46cccbf0-9925-4708-bd05-119652ca2732" (UID: "46cccbf0-9925-4708-bd05-119652ca2732"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.049372 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.049421 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.049447 4829 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46cccbf0-9925-4708-bd05-119652ca2732-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.049478 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6jsw\" (UniqueName: \"kubernetes.io/projected/46cccbf0-9925-4708-bd05-119652ca2732-kube-api-access-g6jsw\") on node \"crc\" DevicePath \"\"" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.188457 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" event={"ID":"46cccbf0-9925-4708-bd05-119652ca2732","Type":"ContainerDied","Data":"c7a679d3bf8232b49518ad312f1ee5a417457987775a57d254111f05020009a8"} Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.188525 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7a679d3bf8232b49518ad312f1ee5a417457987775a57d254111f05020009a8" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.188549 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.321533 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp"] Oct 02 07:42:55 crc kubenswrapper[4829]: E1002 07:42:55.322107 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="extract-content" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.322130 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="extract-content" Oct 02 07:42:55 crc kubenswrapper[4829]: E1002 07:42:55.322161 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="extract-utilities" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.322173 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="extract-utilities" Oct 02 07:42:55 crc kubenswrapper[4829]: E1002 07:42:55.322191 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="registry-server" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.322199 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="registry-server" Oct 02 07:42:55 crc kubenswrapper[4829]: E1002 07:42:55.322223 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46cccbf0-9925-4708-bd05-119652ca2732" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.322318 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="46cccbf0-9925-4708-bd05-119652ca2732" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.322572 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ff0b4a-f70b-4cba-9791-74dd0182aa54" containerName="registry-server" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.322608 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="46cccbf0-9925-4708-bd05-119652ca2732" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.323454 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.326187 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.326405 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.326457 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.327592 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.349846 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp"] Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.355442 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkw22\" (UniqueName: \"kubernetes.io/projected/ac409859-eb90-422d-ad2f-4cf60f171844-kube-api-access-gkw22\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.355577 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.355724 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.457295 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkw22\" (UniqueName: \"kubernetes.io/projected/ac409859-eb90-422d-ad2f-4cf60f171844-kube-api-access-gkw22\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.457351 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.457403 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.462644 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.473537 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.485802 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkw22\" (UniqueName: \"kubernetes.io/projected/ac409859-eb90-422d-ad2f-4cf60f171844-kube-api-access-gkw22\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:55 crc kubenswrapper[4829]: I1002 07:42:55.687117 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:42:56 crc kubenswrapper[4829]: I1002 07:42:56.310872 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:42:56 crc kubenswrapper[4829]: I1002 07:42:56.317858 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp"] Oct 02 07:42:57 crc kubenswrapper[4829]: I1002 07:42:57.222998 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" event={"ID":"ac409859-eb90-422d-ad2f-4cf60f171844","Type":"ContainerStarted","Data":"243fb86b017988326fa28bb7e198d506b91d572070e32673191fc79aa7a318c7"} Oct 02 07:42:57 crc kubenswrapper[4829]: I1002 07:42:57.223781 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" event={"ID":"ac409859-eb90-422d-ad2f-4cf60f171844","Type":"ContainerStarted","Data":"fa1435fbd932a3dd058553399335c2d65f2fe0e3bdbd961b00aa68ec002186b1"} Oct 02 07:42:57 crc kubenswrapper[4829]: I1002 07:42:57.250434 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" podStartSLOduration=1.647434519 podStartE2EDuration="2.250407318s" podCreationTimestamp="2025-10-02 07:42:55 +0000 UTC" firstStartedPulling="2025-10-02 07:42:56.310626163 +0000 UTC m=+1567.650274568" lastFinishedPulling="2025-10-02 07:42:56.913598962 +0000 UTC m=+1568.253247367" observedRunningTime="2025-10-02 07:42:57.240836093 +0000 UTC m=+1568.580484538" watchObservedRunningTime="2025-10-02 07:42:57.250407318 +0000 UTC m=+1568.590055753" Oct 02 07:43:02 crc kubenswrapper[4829]: I1002 07:43:02.461638 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:43:02 crc kubenswrapper[4829]: E1002 07:43:02.462584 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:43:16 crc kubenswrapper[4829]: I1002 07:43:16.461343 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:43:16 crc kubenswrapper[4829]: E1002 07:43:16.461988 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:43:28 crc kubenswrapper[4829]: I1002 07:43:28.053841 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-lh5bj"] Oct 02 07:43:28 crc kubenswrapper[4829]: I1002 07:43:28.061951 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-lh5bj"] Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.037898 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-lvvhk"] Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.046015 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-6wckj"] Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.059253 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-lvvhk"] Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.066953 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-6wckj"] Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.469608 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:43:29 crc kubenswrapper[4829]: E1002 07:43:29.470053 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.482200 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24271dcb-9076-48b4-9698-a3e97a4bd625" path="/var/lib/kubelet/pods/24271dcb-9076-48b4-9698-a3e97a4bd625/volumes" Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.483394 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9059dfc6-81c9-4224-b364-94d9d2f50d67" path="/var/lib/kubelet/pods/9059dfc6-81c9-4224-b364-94d9d2f50d67/volumes" Oct 02 07:43:29 crc kubenswrapper[4829]: I1002 07:43:29.485060 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c14ab4eb-2e4c-4196-8689-c58e5c6e4a42" path="/var/lib/kubelet/pods/c14ab4eb-2e4c-4196-8689-c58e5c6e4a42/volumes" Oct 02 07:43:32 crc kubenswrapper[4829]: I1002 07:43:32.041650 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-kc6lg"] Oct 02 07:43:32 crc kubenswrapper[4829]: I1002 07:43:32.057350 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-kc6lg"] Oct 02 07:43:33 crc kubenswrapper[4829]: I1002 07:43:33.484087 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38a3be44-070a-4d1b-acec-d3d7baac2615" path="/var/lib/kubelet/pods/38a3be44-070a-4d1b-acec-d3d7baac2615/volumes" Oct 02 07:43:38 crc kubenswrapper[4829]: I1002 07:43:38.059719 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3b4a-account-create-w6wq5"] Oct 02 07:43:38 crc kubenswrapper[4829]: I1002 07:43:38.076208 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3b4a-account-create-w6wq5"] Oct 02 07:43:39 crc kubenswrapper[4829]: I1002 07:43:39.034798 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-4939-account-create-2wx9l"] Oct 02 07:43:39 crc kubenswrapper[4829]: I1002 07:43:39.045706 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-4939-account-create-2wx9l"] Oct 02 07:43:39 crc kubenswrapper[4829]: I1002 07:43:39.477891 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ae65aa7-5414-49a2-bd11-dd2209873f7e" path="/var/lib/kubelet/pods/9ae65aa7-5414-49a2-bd11-dd2209873f7e/volumes" Oct 02 07:43:39 crc kubenswrapper[4829]: I1002 07:43:39.479420 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac2d616a-91bd-4aa2-b818-09bd73c59320" path="/var/lib/kubelet/pods/ac2d616a-91bd-4aa2-b818-09bd73c59320/volumes" Oct 02 07:43:40 crc kubenswrapper[4829]: I1002 07:43:40.462339 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:43:40 crc kubenswrapper[4829]: E1002 07:43:40.463080 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:43:51 crc kubenswrapper[4829]: I1002 07:43:51.461658 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:43:51 crc kubenswrapper[4829]: E1002 07:43:51.463108 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:43:55 crc kubenswrapper[4829]: I1002 07:43:55.030893 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-6628-account-create-dwgd4"] Oct 02 07:43:55 crc kubenswrapper[4829]: I1002 07:43:55.039098 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-6628-account-create-dwgd4"] Oct 02 07:43:55 crc kubenswrapper[4829]: I1002 07:43:55.481329 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3fba779-519a-4959-860b-e6daf434d2ee" path="/var/lib/kubelet/pods/e3fba779-519a-4959-860b-e6daf434d2ee/volumes" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.703791 4829 scope.go:117] "RemoveContainer" containerID="1af76466372c9c037930e400e2270fe1e220f18d8c50a01d7806b3f450cf6669" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.737857 4829 scope.go:117] "RemoveContainer" containerID="83f4ac4c30cc9d4bbb7770e1caa7034916ee286611cc93005538f68abe316fe2" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.788442 4829 scope.go:117] "RemoveContainer" containerID="ed6cef3818f9f4fcdbee3cf10d168ef0310b0257286cf79c3d92c6bbe8663bbd" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.831486 4829 scope.go:117] "RemoveContainer" containerID="15af294706718995e7df334a59da66bc1211326e23ae46f9db15f14da5ce4311" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.875890 4829 scope.go:117] "RemoveContainer" containerID="34a54c0c6302c4c359abf36f5599e8ea69273606ea5a3310f8eeca7f4eb10b1d" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.917395 4829 scope.go:117] "RemoveContainer" containerID="7f348771e01df8391a5d48147093da9cd70eda5bd6b1bbb8a86a7bd21cdb06ba" Oct 02 07:43:56 crc kubenswrapper[4829]: I1002 07:43:56.955327 4829 scope.go:117] "RemoveContainer" containerID="2a9b9f8b92b38426762f0678d8ba976f47209d1139aa56ea90af2247985aece8" Oct 02 07:43:57 crc kubenswrapper[4829]: I1002 07:43:57.032984 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-zrcg4"] Oct 02 07:43:57 crc kubenswrapper[4829]: I1002 07:43:57.041038 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-zrcg4"] Oct 02 07:43:57 crc kubenswrapper[4829]: I1002 07:43:57.480603 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61eaa9fd-41d9-4341-8f74-acef1d2782d6" path="/var/lib/kubelet/pods/61eaa9fd-41d9-4341-8f74-acef1d2782d6/volumes" Oct 02 07:43:58 crc kubenswrapper[4829]: I1002 07:43:58.032593 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-z22p2"] Oct 02 07:43:58 crc kubenswrapper[4829]: I1002 07:43:58.052356 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-s25rw"] Oct 02 07:43:58 crc kubenswrapper[4829]: I1002 07:43:58.062645 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-z22p2"] Oct 02 07:43:58 crc kubenswrapper[4829]: I1002 07:43:58.074100 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-s25rw"] Oct 02 07:43:59 crc kubenswrapper[4829]: I1002 07:43:59.035845 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db66-account-create-ldcv7"] Oct 02 07:43:59 crc kubenswrapper[4829]: I1002 07:43:59.050995 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db66-account-create-ldcv7"] Oct 02 07:43:59 crc kubenswrapper[4829]: I1002 07:43:59.478746 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e9830ff-3291-43f6-89cf-3911036a3350" path="/var/lib/kubelet/pods/6e9830ff-3291-43f6-89cf-3911036a3350/volumes" Oct 02 07:43:59 crc kubenswrapper[4829]: I1002 07:43:59.480840 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9df948cb-5fcf-4d8c-868d-4d80cbe4c172" path="/var/lib/kubelet/pods/9df948cb-5fcf-4d8c-868d-4d80cbe4c172/volumes" Oct 02 07:43:59 crc kubenswrapper[4829]: I1002 07:43:59.481883 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c61c72f6-3856-4475-9d67-eaa307c9d23e" path="/var/lib/kubelet/pods/c61c72f6-3856-4475-9d67-eaa307c9d23e/volumes" Oct 02 07:44:06 crc kubenswrapper[4829]: I1002 07:44:06.461599 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:44:06 crc kubenswrapper[4829]: E1002 07:44:06.463114 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:44:08 crc kubenswrapper[4829]: I1002 07:44:08.054766 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-676c-account-create-6wfbq"] Oct 02 07:44:08 crc kubenswrapper[4829]: I1002 07:44:08.068907 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-b5txm"] Oct 02 07:44:08 crc kubenswrapper[4829]: I1002 07:44:08.082684 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-b5txm"] Oct 02 07:44:08 crc kubenswrapper[4829]: I1002 07:44:08.096855 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-676c-account-create-6wfbq"] Oct 02 07:44:09 crc kubenswrapper[4829]: I1002 07:44:09.027124 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-3056-account-create-gdp9h"] Oct 02 07:44:09 crc kubenswrapper[4829]: I1002 07:44:09.035058 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-3056-account-create-gdp9h"] Oct 02 07:44:09 crc kubenswrapper[4829]: I1002 07:44:09.475475 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f1fa854-ffef-4cfe-a58f-7fc7f382af72" path="/var/lib/kubelet/pods/0f1fa854-ffef-4cfe-a58f-7fc7f382af72/volumes" Oct 02 07:44:09 crc kubenswrapper[4829]: I1002 07:44:09.476855 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97b06d08-abdf-4107-9561-9b81768e183c" path="/var/lib/kubelet/pods/97b06d08-abdf-4107-9561-9b81768e183c/volumes" Oct 02 07:44:09 crc kubenswrapper[4829]: I1002 07:44:09.477908 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebc9e693-09fe-4577-a32b-cad2af43ab8a" path="/var/lib/kubelet/pods/ebc9e693-09fe-4577-a32b-cad2af43ab8a/volumes" Oct 02 07:44:17 crc kubenswrapper[4829]: I1002 07:44:17.462192 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:44:17 crc kubenswrapper[4829]: E1002 07:44:17.463552 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:44:20 crc kubenswrapper[4829]: I1002 07:44:20.033185 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8fcd-account-create-7fhc4"] Oct 02 07:44:20 crc kubenswrapper[4829]: I1002 07:44:20.042806 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8fcd-account-create-7fhc4"] Oct 02 07:44:21 crc kubenswrapper[4829]: I1002 07:44:21.030703 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-89hnk"] Oct 02 07:44:21 crc kubenswrapper[4829]: I1002 07:44:21.040179 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-89hnk"] Oct 02 07:44:21 crc kubenswrapper[4829]: I1002 07:44:21.478773 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="971bceb7-6e70-4502-b0bc-f5071f3998ea" path="/var/lib/kubelet/pods/971bceb7-6e70-4502-b0bc-f5071f3998ea/volumes" Oct 02 07:44:21 crc kubenswrapper[4829]: I1002 07:44:21.480547 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98649236-1856-4a42-8171-9435997ecb47" path="/var/lib/kubelet/pods/98649236-1856-4a42-8171-9435997ecb47/volumes" Oct 02 07:44:31 crc kubenswrapper[4829]: I1002 07:44:31.460752 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:44:31 crc kubenswrapper[4829]: E1002 07:44:31.461975 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:44:41 crc kubenswrapper[4829]: I1002 07:44:41.481420 4829 generic.go:334] "Generic (PLEG): container finished" podID="ac409859-eb90-422d-ad2f-4cf60f171844" containerID="243fb86b017988326fa28bb7e198d506b91d572070e32673191fc79aa7a318c7" exitCode=0 Oct 02 07:44:41 crc kubenswrapper[4829]: I1002 07:44:41.481496 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" event={"ID":"ac409859-eb90-422d-ad2f-4cf60f171844","Type":"ContainerDied","Data":"243fb86b017988326fa28bb7e198d506b91d572070e32673191fc79aa7a318c7"} Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.059074 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.142876 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-ssh-key\") pod \"ac409859-eb90-422d-ad2f-4cf60f171844\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.143028 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-inventory\") pod \"ac409859-eb90-422d-ad2f-4cf60f171844\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.143081 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkw22\" (UniqueName: \"kubernetes.io/projected/ac409859-eb90-422d-ad2f-4cf60f171844-kube-api-access-gkw22\") pod \"ac409859-eb90-422d-ad2f-4cf60f171844\" (UID: \"ac409859-eb90-422d-ad2f-4cf60f171844\") " Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.150756 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac409859-eb90-422d-ad2f-4cf60f171844-kube-api-access-gkw22" (OuterVolumeSpecName: "kube-api-access-gkw22") pod "ac409859-eb90-422d-ad2f-4cf60f171844" (UID: "ac409859-eb90-422d-ad2f-4cf60f171844"). InnerVolumeSpecName "kube-api-access-gkw22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.179380 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-inventory" (OuterVolumeSpecName: "inventory") pod "ac409859-eb90-422d-ad2f-4cf60f171844" (UID: "ac409859-eb90-422d-ad2f-4cf60f171844"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.183604 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac409859-eb90-422d-ad2f-4cf60f171844" (UID: "ac409859-eb90-422d-ad2f-4cf60f171844"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.245731 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.245787 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkw22\" (UniqueName: \"kubernetes.io/projected/ac409859-eb90-422d-ad2f-4cf60f171844-kube-api-access-gkw22\") on node \"crc\" DevicePath \"\"" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.245806 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac409859-eb90-422d-ad2f-4cf60f171844-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.514919 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" event={"ID":"ac409859-eb90-422d-ad2f-4cf60f171844","Type":"ContainerDied","Data":"fa1435fbd932a3dd058553399335c2d65f2fe0e3bdbd961b00aa68ec002186b1"} Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.514973 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa1435fbd932a3dd058553399335c2d65f2fe0e3bdbd961b00aa68ec002186b1" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.515049 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.647105 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz"] Oct 02 07:44:43 crc kubenswrapper[4829]: E1002 07:44:43.647694 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac409859-eb90-422d-ad2f-4cf60f171844" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.647712 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac409859-eb90-422d-ad2f-4cf60f171844" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.647911 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac409859-eb90-422d-ad2f-4cf60f171844" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.648607 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.651730 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.652516 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.652660 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.652784 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.669429 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz"] Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.758762 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75xcv\" (UniqueName: \"kubernetes.io/projected/e6b2c2b0-9a68-4525-9f18-64b8e4354246-kube-api-access-75xcv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.758822 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.759089 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.861384 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75xcv\" (UniqueName: \"kubernetes.io/projected/e6b2c2b0-9a68-4525-9f18-64b8e4354246-kube-api-access-75xcv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.861921 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.862272 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.868331 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.873684 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.884376 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75xcv\" (UniqueName: \"kubernetes.io/projected/e6b2c2b0-9a68-4525-9f18-64b8e4354246-kube-api-access-75xcv\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:43 crc kubenswrapper[4829]: I1002 07:44:43.983564 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:44:44 crc kubenswrapper[4829]: I1002 07:44:44.353839 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz"] Oct 02 07:44:44 crc kubenswrapper[4829]: I1002 07:44:44.528764 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" event={"ID":"e6b2c2b0-9a68-4525-9f18-64b8e4354246","Type":"ContainerStarted","Data":"9003241f7ce36c51fb11a88827f55aa1d38ecb1ba0c732c2d9173209214a192c"} Oct 02 07:44:45 crc kubenswrapper[4829]: I1002 07:44:45.549989 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" event={"ID":"e6b2c2b0-9a68-4525-9f18-64b8e4354246","Type":"ContainerStarted","Data":"e17ea8a28fc9d46d18e840c410b53e2e0d63455cd065bbed012353451e925a88"} Oct 02 07:44:45 crc kubenswrapper[4829]: I1002 07:44:45.572381 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" podStartSLOduration=1.8573307890000001 podStartE2EDuration="2.57236269s" podCreationTimestamp="2025-10-02 07:44:43 +0000 UTC" firstStartedPulling="2025-10-02 07:44:44.359902753 +0000 UTC m=+1675.699551158" lastFinishedPulling="2025-10-02 07:44:45.074934654 +0000 UTC m=+1676.414583059" observedRunningTime="2025-10-02 07:44:45.567758673 +0000 UTC m=+1676.907407098" watchObservedRunningTime="2025-10-02 07:44:45.57236269 +0000 UTC m=+1676.912011105" Oct 02 07:44:46 crc kubenswrapper[4829]: I1002 07:44:46.461575 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:44:46 crc kubenswrapper[4829]: E1002 07:44:46.462270 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:44:50 crc kubenswrapper[4829]: I1002 07:44:50.067738 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-tzrbj"] Oct 02 07:44:50 crc kubenswrapper[4829]: I1002 07:44:50.078465 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-tzrbj"] Oct 02 07:44:51 crc kubenswrapper[4829]: I1002 07:44:51.476295 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bd0e965-1cf0-4edd-856e-d424c91d80a9" path="/var/lib/kubelet/pods/8bd0e965-1cf0-4edd-856e-d424c91d80a9/volumes" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.047252 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-28kmg"] Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.066833 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-28kmg"] Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.150081 4829 scope.go:117] "RemoveContainer" containerID="17ede04b6f9b8448f6085cd42c78d9aedd5c585a76ef015f6cf4aecbed8a0c76" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.176432 4829 scope.go:117] "RemoveContainer" containerID="71f7a1ddbf5b094e2d3daf3cf8edf4985a804ce1b17904051424458e37f4321c" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.238178 4829 scope.go:117] "RemoveContainer" containerID="484e8753483fb11a7420eb11926e3496f314702908fcb1ee6f97d0fe9c517ce9" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.293988 4829 scope.go:117] "RemoveContainer" containerID="32cad8b33fedf01de6a1920e2cf46d92655ab50de25f5d22435a741bb4077496" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.346951 4829 scope.go:117] "RemoveContainer" containerID="bf6db1a9b6aedd6448837ea1085c40b0a89f07480ad31e3dc8203c881da5fe5e" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.438397 4829 scope.go:117] "RemoveContainer" containerID="1aa8fb8ff73f092bc301d8f6acdd4500b47f9ce3b4bdc9d83539574ff1a72b76" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.480931 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af15cc1d-88dc-44f9-b43c-9570dfc4fd65" path="/var/lib/kubelet/pods/af15cc1d-88dc-44f9-b43c-9570dfc4fd65/volumes" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.488508 4829 scope.go:117] "RemoveContainer" containerID="51fae9a9d6dde21c2c6b451d316fee6b2f43669963e47a89ea91d44ec5510306" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.515761 4829 scope.go:117] "RemoveContainer" containerID="6a1b99e31061ef8934d07fedcbaf6f21a33eef475ba6b90c1337e3e531d8d40f" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.543591 4829 scope.go:117] "RemoveContainer" containerID="2ee0d197c56046778ed4d28f0d37be24fddd763c1bbfb976c55c42dc784374ef" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.570558 4829 scope.go:117] "RemoveContainer" containerID="7c06500225e2b9e067576af209242ded3b8902bc6166876744edd7adb866dcad" Oct 02 07:44:57 crc kubenswrapper[4829]: I1002 07:44:57.611786 4829 scope.go:117] "RemoveContainer" containerID="a2af9b1d63b65d5d17cd15749891c841dc3fa078036044911f30035e6fbf5489" Oct 02 07:44:58 crc kubenswrapper[4829]: I1002 07:44:58.462070 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:44:58 crc kubenswrapper[4829]: E1002 07:44:58.463032 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:44:59 crc kubenswrapper[4829]: I1002 07:44:59.038755 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-8tjd9"] Oct 02 07:44:59 crc kubenswrapper[4829]: I1002 07:44:59.052084 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-g5dzf"] Oct 02 07:44:59 crc kubenswrapper[4829]: I1002 07:44:59.060064 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-8tjd9"] Oct 02 07:44:59 crc kubenswrapper[4829]: I1002 07:44:59.067278 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-g5dzf"] Oct 02 07:44:59 crc kubenswrapper[4829]: I1002 07:44:59.484707 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8adee8c4-a6f1-42cb-ae32-a7cc57b911b4" path="/var/lib/kubelet/pods/8adee8c4-a6f1-42cb-ae32-a7cc57b911b4/volumes" Oct 02 07:44:59 crc kubenswrapper[4829]: I1002 07:44:59.486111 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43b663f-cfca-459a-b513-ed16fdecd2f6" path="/var/lib/kubelet/pods/e43b663f-cfca-459a-b513-ed16fdecd2f6/volumes" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.149708 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw"] Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.151327 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.154646 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.157726 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.170052 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw"] Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.231155 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686nl\" (UniqueName: \"kubernetes.io/projected/9b5d0a66-00ea-4f01-9040-6b53de05c254-kube-api-access-686nl\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.231628 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b5d0a66-00ea-4f01-9040-6b53de05c254-secret-volume\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.231684 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b5d0a66-00ea-4f01-9040-6b53de05c254-config-volume\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.333101 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686nl\" (UniqueName: \"kubernetes.io/projected/9b5d0a66-00ea-4f01-9040-6b53de05c254-kube-api-access-686nl\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.333216 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b5d0a66-00ea-4f01-9040-6b53de05c254-secret-volume\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.333268 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b5d0a66-00ea-4f01-9040-6b53de05c254-config-volume\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.334148 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b5d0a66-00ea-4f01-9040-6b53de05c254-config-volume\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.341577 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b5d0a66-00ea-4f01-9040-6b53de05c254-secret-volume\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.353429 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686nl\" (UniqueName: \"kubernetes.io/projected/9b5d0a66-00ea-4f01-9040-6b53de05c254-kube-api-access-686nl\") pod \"collect-profiles-29323185-6hnnw\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.493607 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:00 crc kubenswrapper[4829]: I1002 07:45:00.927763 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw"] Oct 02 07:45:00 crc kubenswrapper[4829]: W1002 07:45:00.939837 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b5d0a66_00ea_4f01_9040_6b53de05c254.slice/crio-475a3b710303282b672593b87e1ddf33c7001f0693adfce4ed8156d752f24caa WatchSource:0}: Error finding container 475a3b710303282b672593b87e1ddf33c7001f0693adfce4ed8156d752f24caa: Status 404 returned error can't find the container with id 475a3b710303282b672593b87e1ddf33c7001f0693adfce4ed8156d752f24caa Oct 02 07:45:01 crc kubenswrapper[4829]: I1002 07:45:01.753048 4829 generic.go:334] "Generic (PLEG): container finished" podID="9b5d0a66-00ea-4f01-9040-6b53de05c254" containerID="198a1730723e0f0dfd27f4ff4b99c4bac6cc7699d97cabad8e606d84ffeb7de4" exitCode=0 Oct 02 07:45:01 crc kubenswrapper[4829]: I1002 07:45:01.753448 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" event={"ID":"9b5d0a66-00ea-4f01-9040-6b53de05c254","Type":"ContainerDied","Data":"198a1730723e0f0dfd27f4ff4b99c4bac6cc7699d97cabad8e606d84ffeb7de4"} Oct 02 07:45:01 crc kubenswrapper[4829]: I1002 07:45:01.753495 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" event={"ID":"9b5d0a66-00ea-4f01-9040-6b53de05c254","Type":"ContainerStarted","Data":"475a3b710303282b672593b87e1ddf33c7001f0693adfce4ed8156d752f24caa"} Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.226316 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.395741 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b5d0a66-00ea-4f01-9040-6b53de05c254-secret-volume\") pod \"9b5d0a66-00ea-4f01-9040-6b53de05c254\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.395896 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-686nl\" (UniqueName: \"kubernetes.io/projected/9b5d0a66-00ea-4f01-9040-6b53de05c254-kube-api-access-686nl\") pod \"9b5d0a66-00ea-4f01-9040-6b53de05c254\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.395994 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b5d0a66-00ea-4f01-9040-6b53de05c254-config-volume\") pod \"9b5d0a66-00ea-4f01-9040-6b53de05c254\" (UID: \"9b5d0a66-00ea-4f01-9040-6b53de05c254\") " Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.396927 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b5d0a66-00ea-4f01-9040-6b53de05c254-config-volume" (OuterVolumeSpecName: "config-volume") pod "9b5d0a66-00ea-4f01-9040-6b53de05c254" (UID: "9b5d0a66-00ea-4f01-9040-6b53de05c254"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.402487 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5d0a66-00ea-4f01-9040-6b53de05c254-kube-api-access-686nl" (OuterVolumeSpecName: "kube-api-access-686nl") pod "9b5d0a66-00ea-4f01-9040-6b53de05c254" (UID: "9b5d0a66-00ea-4f01-9040-6b53de05c254"). InnerVolumeSpecName "kube-api-access-686nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.402559 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5d0a66-00ea-4f01-9040-6b53de05c254-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9b5d0a66-00ea-4f01-9040-6b53de05c254" (UID: "9b5d0a66-00ea-4f01-9040-6b53de05c254"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.499183 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9b5d0a66-00ea-4f01-9040-6b53de05c254-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.499279 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9b5d0a66-00ea-4f01-9040-6b53de05c254-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.499301 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-686nl\" (UniqueName: \"kubernetes.io/projected/9b5d0a66-00ea-4f01-9040-6b53de05c254-kube-api-access-686nl\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.783293 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" event={"ID":"9b5d0a66-00ea-4f01-9040-6b53de05c254","Type":"ContainerDied","Data":"475a3b710303282b672593b87e1ddf33c7001f0693adfce4ed8156d752f24caa"} Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.783361 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="475a3b710303282b672593b87e1ddf33c7001f0693adfce4ed8156d752f24caa" Oct 02 07:45:03 crc kubenswrapper[4829]: I1002 07:45:03.783386 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw" Oct 02 07:45:11 crc kubenswrapper[4829]: I1002 07:45:11.461096 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:45:11 crc kubenswrapper[4829]: E1002 07:45:11.462109 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:45:16 crc kubenswrapper[4829]: I1002 07:45:16.069403 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dtlnv"] Oct 02 07:45:16 crc kubenswrapper[4829]: I1002 07:45:16.080603 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xhn5v"] Oct 02 07:45:16 crc kubenswrapper[4829]: I1002 07:45:16.092596 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dtlnv"] Oct 02 07:45:16 crc kubenswrapper[4829]: I1002 07:45:16.103069 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xhn5v"] Oct 02 07:45:17 crc kubenswrapper[4829]: I1002 07:45:17.472870 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b3627e8-07df-4d85-a723-a763eb14c52b" path="/var/lib/kubelet/pods/1b3627e8-07df-4d85-a723-a763eb14c52b/volumes" Oct 02 07:45:17 crc kubenswrapper[4829]: I1002 07:45:17.474072 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea5c0100-6e79-43fb-9761-62c7e758e891" path="/var/lib/kubelet/pods/ea5c0100-6e79-43fb-9761-62c7e758e891/volumes" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.687333 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-br8b2"] Oct 02 07:45:20 crc kubenswrapper[4829]: E1002 07:45:20.688638 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5d0a66-00ea-4f01-9040-6b53de05c254" containerName="collect-profiles" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.688660 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5d0a66-00ea-4f01-9040-6b53de05c254" containerName="collect-profiles" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.688986 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5d0a66-00ea-4f01-9040-6b53de05c254" containerName="collect-profiles" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.691644 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.698130 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-br8b2"] Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.877568 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s5km\" (UniqueName: \"kubernetes.io/projected/a796d106-1f58-418d-8dbb-5172dce0db93-kube-api-access-2s5km\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.877820 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-catalog-content\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.877888 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-utilities\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.879350 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hv2nx"] Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.881238 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.903005 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hv2nx"] Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980153 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-catalog-content\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980293 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-utilities\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980370 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-utilities\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980480 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-catalog-content\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980579 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s5km\" (UniqueName: \"kubernetes.io/projected/a796d106-1f58-418d-8dbb-5172dce0db93-kube-api-access-2s5km\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980664 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-php8n\" (UniqueName: \"kubernetes.io/projected/af87ca08-b9dd-4363-ab52-3f9605c309c2-kube-api-access-php8n\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980683 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-catalog-content\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:20 crc kubenswrapper[4829]: I1002 07:45:20.980693 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-utilities\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.002686 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s5km\" (UniqueName: \"kubernetes.io/projected/a796d106-1f58-418d-8dbb-5172dce0db93-kube-api-access-2s5km\") pod \"certified-operators-br8b2\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.018409 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.081440 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-utilities\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.081511 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-catalog-content\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.081563 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-php8n\" (UniqueName: \"kubernetes.io/projected/af87ca08-b9dd-4363-ab52-3f9605c309c2-kube-api-access-php8n\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.082378 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-utilities\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.082421 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-catalog-content\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.100814 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-php8n\" (UniqueName: \"kubernetes.io/projected/af87ca08-b9dd-4363-ab52-3f9605c309c2-kube-api-access-php8n\") pod \"community-operators-hv2nx\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.203666 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.512821 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-br8b2"] Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.643931 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hv2nx"] Oct 02 07:45:21 crc kubenswrapper[4829]: W1002 07:45:21.653132 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf87ca08_b9dd_4363_ab52_3f9605c309c2.slice/crio-1b8e6b5e1280a49b7cd2c6468a01c0eaefd89be574c5882320faee5bd139aefe WatchSource:0}: Error finding container 1b8e6b5e1280a49b7cd2c6468a01c0eaefd89be574c5882320faee5bd139aefe: Status 404 returned error can't find the container with id 1b8e6b5e1280a49b7cd2c6468a01c0eaefd89be574c5882320faee5bd139aefe Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.975965 4829 generic.go:334] "Generic (PLEG): container finished" podID="a796d106-1f58-418d-8dbb-5172dce0db93" containerID="48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c" exitCode=0 Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.976006 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerDied","Data":"48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c"} Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.976438 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerStarted","Data":"1d234ce90c3645daea7e7470f91b611952b5b8f10bbb7fe359afc2930d581ad2"} Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.978199 4829 generic.go:334] "Generic (PLEG): container finished" podID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerID="bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317" exitCode=0 Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.978270 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerDied","Data":"bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317"} Oct 02 07:45:21 crc kubenswrapper[4829]: I1002 07:45:21.978352 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerStarted","Data":"1b8e6b5e1280a49b7cd2c6468a01c0eaefd89be574c5882320faee5bd139aefe"} Oct 02 07:45:22 crc kubenswrapper[4829]: I1002 07:45:22.461518 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:45:22 crc kubenswrapper[4829]: E1002 07:45:22.462258 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:45:22 crc kubenswrapper[4829]: I1002 07:45:22.988733 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerStarted","Data":"f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170"} Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.080724 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dkrcj"] Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.083465 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.094473 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkrcj"] Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.221251 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-utilities\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.221313 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v665m\" (UniqueName: \"kubernetes.io/projected/1e082a97-b508-49a7-86d2-57c2453dd7b8-kube-api-access-v665m\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.221730 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-catalog-content\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.323163 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-utilities\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.323283 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v665m\" (UniqueName: \"kubernetes.io/projected/1e082a97-b508-49a7-86d2-57c2453dd7b8-kube-api-access-v665m\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.323381 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-catalog-content\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.323681 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-utilities\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.323819 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-catalog-content\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.344627 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v665m\" (UniqueName: \"kubernetes.io/projected/1e082a97-b508-49a7-86d2-57c2453dd7b8-kube-api-access-v665m\") pod \"redhat-marketplace-dkrcj\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.405712 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:23 crc kubenswrapper[4829]: I1002 07:45:23.754609 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkrcj"] Oct 02 07:45:23 crc kubenswrapper[4829]: W1002 07:45:23.760440 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e082a97_b508_49a7_86d2_57c2453dd7b8.slice/crio-7678186818a1c99f10daa350b209b4cd8c50fee726a8e1f58cc2c781e641ac4a WatchSource:0}: Error finding container 7678186818a1c99f10daa350b209b4cd8c50fee726a8e1f58cc2c781e641ac4a: Status 404 returned error can't find the container with id 7678186818a1c99f10daa350b209b4cd8c50fee726a8e1f58cc2c781e641ac4a Oct 02 07:45:24 crc kubenswrapper[4829]: I1002 07:45:24.002298 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkrcj" event={"ID":"1e082a97-b508-49a7-86d2-57c2453dd7b8","Type":"ContainerStarted","Data":"7678186818a1c99f10daa350b209b4cd8c50fee726a8e1f58cc2c781e641ac4a"} Oct 02 07:45:24 crc kubenswrapper[4829]: I1002 07:45:24.004990 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerStarted","Data":"09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd"} Oct 02 07:45:25 crc kubenswrapper[4829]: I1002 07:45:25.017274 4829 generic.go:334] "Generic (PLEG): container finished" podID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerID="1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832" exitCode=0 Oct 02 07:45:25 crc kubenswrapper[4829]: I1002 07:45:25.017459 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkrcj" event={"ID":"1e082a97-b508-49a7-86d2-57c2453dd7b8","Type":"ContainerDied","Data":"1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832"} Oct 02 07:45:26 crc kubenswrapper[4829]: I1002 07:45:26.028979 4829 generic.go:334] "Generic (PLEG): container finished" podID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerID="5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8" exitCode=0 Oct 02 07:45:26 crc kubenswrapper[4829]: I1002 07:45:26.029184 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkrcj" event={"ID":"1e082a97-b508-49a7-86d2-57c2453dd7b8","Type":"ContainerDied","Data":"5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8"} Oct 02 07:45:26 crc kubenswrapper[4829]: I1002 07:45:26.033531 4829 generic.go:334] "Generic (PLEG): container finished" podID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerID="09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd" exitCode=0 Oct 02 07:45:26 crc kubenswrapper[4829]: I1002 07:45:26.033620 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerDied","Data":"09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd"} Oct 02 07:45:26 crc kubenswrapper[4829]: I1002 07:45:26.037434 4829 generic.go:334] "Generic (PLEG): container finished" podID="a796d106-1f58-418d-8dbb-5172dce0db93" containerID="f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170" exitCode=0 Oct 02 07:45:26 crc kubenswrapper[4829]: I1002 07:45:26.037473 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerDied","Data":"f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170"} Oct 02 07:45:27 crc kubenswrapper[4829]: I1002 07:45:27.066610 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerStarted","Data":"e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196"} Oct 02 07:45:27 crc kubenswrapper[4829]: I1002 07:45:27.072789 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkrcj" event={"ID":"1e082a97-b508-49a7-86d2-57c2453dd7b8","Type":"ContainerStarted","Data":"381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028"} Oct 02 07:45:27 crc kubenswrapper[4829]: I1002 07:45:27.075782 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerStarted","Data":"13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9"} Oct 02 07:45:27 crc kubenswrapper[4829]: I1002 07:45:27.096015 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-br8b2" podStartSLOduration=2.584359748 podStartE2EDuration="7.096002289s" podCreationTimestamp="2025-10-02 07:45:20 +0000 UTC" firstStartedPulling="2025-10-02 07:45:21.978248208 +0000 UTC m=+1713.317896623" lastFinishedPulling="2025-10-02 07:45:26.489890749 +0000 UTC m=+1717.829539164" observedRunningTime="2025-10-02 07:45:27.092979065 +0000 UTC m=+1718.432627470" watchObservedRunningTime="2025-10-02 07:45:27.096002289 +0000 UTC m=+1718.435650694" Oct 02 07:45:27 crc kubenswrapper[4829]: I1002 07:45:27.115748 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hv2nx" podStartSLOduration=2.477107554 podStartE2EDuration="7.115732404s" podCreationTimestamp="2025-10-02 07:45:20 +0000 UTC" firstStartedPulling="2025-10-02 07:45:21.980110267 +0000 UTC m=+1713.319758682" lastFinishedPulling="2025-10-02 07:45:26.618735117 +0000 UTC m=+1717.958383532" observedRunningTime="2025-10-02 07:45:27.115299591 +0000 UTC m=+1718.454948006" watchObservedRunningTime="2025-10-02 07:45:27.115732404 +0000 UTC m=+1718.455380799" Oct 02 07:45:27 crc kubenswrapper[4829]: I1002 07:45:27.138937 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dkrcj" podStartSLOduration=2.594405087 podStartE2EDuration="4.138920077s" podCreationTimestamp="2025-10-02 07:45:23 +0000 UTC" firstStartedPulling="2025-10-02 07:45:25.020363707 +0000 UTC m=+1716.360012152" lastFinishedPulling="2025-10-02 07:45:26.564878717 +0000 UTC m=+1717.904527142" observedRunningTime="2025-10-02 07:45:27.135983155 +0000 UTC m=+1718.475631560" watchObservedRunningTime="2025-10-02 07:45:27.138920077 +0000 UTC m=+1718.478568482" Oct 02 07:45:31 crc kubenswrapper[4829]: I1002 07:45:31.018745 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:31 crc kubenswrapper[4829]: I1002 07:45:31.019282 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:31 crc kubenswrapper[4829]: I1002 07:45:31.114468 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:31 crc kubenswrapper[4829]: I1002 07:45:31.165374 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:31 crc kubenswrapper[4829]: I1002 07:45:31.204403 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:31 crc kubenswrapper[4829]: I1002 07:45:31.204537 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:32 crc kubenswrapper[4829]: I1002 07:45:32.254790 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hv2nx" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="registry-server" probeResult="failure" output=< Oct 02 07:45:32 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:45:32 crc kubenswrapper[4829]: > Oct 02 07:45:33 crc kubenswrapper[4829]: I1002 07:45:33.405899 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:33 crc kubenswrapper[4829]: I1002 07:45:33.406404 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:33 crc kubenswrapper[4829]: I1002 07:45:33.453445 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:34 crc kubenswrapper[4829]: I1002 07:45:34.219494 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:34 crc kubenswrapper[4829]: I1002 07:45:34.461709 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:45:34 crc kubenswrapper[4829]: E1002 07:45:34.462465 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.077242 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-br8b2"] Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.077959 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-br8b2" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="registry-server" containerID="cri-o://e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196" gracePeriod=2 Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.517276 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.712285 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-utilities\") pod \"a796d106-1f58-418d-8dbb-5172dce0db93\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.712401 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s5km\" (UniqueName: \"kubernetes.io/projected/a796d106-1f58-418d-8dbb-5172dce0db93-kube-api-access-2s5km\") pod \"a796d106-1f58-418d-8dbb-5172dce0db93\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.712604 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-catalog-content\") pod \"a796d106-1f58-418d-8dbb-5172dce0db93\" (UID: \"a796d106-1f58-418d-8dbb-5172dce0db93\") " Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.713025 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-utilities" (OuterVolumeSpecName: "utilities") pod "a796d106-1f58-418d-8dbb-5172dce0db93" (UID: "a796d106-1f58-418d-8dbb-5172dce0db93"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.713840 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.720274 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a796d106-1f58-418d-8dbb-5172dce0db93-kube-api-access-2s5km" (OuterVolumeSpecName: "kube-api-access-2s5km") pod "a796d106-1f58-418d-8dbb-5172dce0db93" (UID: "a796d106-1f58-418d-8dbb-5172dce0db93"). InnerVolumeSpecName "kube-api-access-2s5km". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.785476 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a796d106-1f58-418d-8dbb-5172dce0db93" (UID: "a796d106-1f58-418d-8dbb-5172dce0db93"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.816613 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a796d106-1f58-418d-8dbb-5172dce0db93-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:36 crc kubenswrapper[4829]: I1002 07:45:36.816649 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s5km\" (UniqueName: \"kubernetes.io/projected/a796d106-1f58-418d-8dbb-5172dce0db93-kube-api-access-2s5km\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.177838 4829 generic.go:334] "Generic (PLEG): container finished" podID="a796d106-1f58-418d-8dbb-5172dce0db93" containerID="e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196" exitCode=0 Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.177922 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-br8b2" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.177956 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerDied","Data":"e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196"} Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.178397 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-br8b2" event={"ID":"a796d106-1f58-418d-8dbb-5172dce0db93","Type":"ContainerDied","Data":"1d234ce90c3645daea7e7470f91b611952b5b8f10bbb7fe359afc2930d581ad2"} Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.178422 4829 scope.go:117] "RemoveContainer" containerID="e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.209081 4829 scope.go:117] "RemoveContainer" containerID="f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.211523 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-br8b2"] Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.219657 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-br8b2"] Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.230032 4829 scope.go:117] "RemoveContainer" containerID="48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.285418 4829 scope.go:117] "RemoveContainer" containerID="e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196" Oct 02 07:45:37 crc kubenswrapper[4829]: E1002 07:45:37.285746 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196\": container with ID starting with e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196 not found: ID does not exist" containerID="e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.285777 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196"} err="failed to get container status \"e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196\": rpc error: code = NotFound desc = could not find container \"e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196\": container with ID starting with e19c6b03a951943a4aafe46d4b6b85f870656669d172dff8f43942232f2d1196 not found: ID does not exist" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.285797 4829 scope.go:117] "RemoveContainer" containerID="f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170" Oct 02 07:45:37 crc kubenswrapper[4829]: E1002 07:45:37.286013 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170\": container with ID starting with f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170 not found: ID does not exist" containerID="f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.286035 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170"} err="failed to get container status \"f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170\": rpc error: code = NotFound desc = could not find container \"f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170\": container with ID starting with f5052cea03e8923b529d28577b6d75b68b4f1fd719732be57515c7992191f170 not found: ID does not exist" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.286047 4829 scope.go:117] "RemoveContainer" containerID="48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c" Oct 02 07:45:37 crc kubenswrapper[4829]: E1002 07:45:37.286278 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c\": container with ID starting with 48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c not found: ID does not exist" containerID="48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.286296 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c"} err="failed to get container status \"48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c\": rpc error: code = NotFound desc = could not find container \"48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c\": container with ID starting with 48a46bc96f646f36ea075cad48bf49ed5689eeaf587e207ecd0bad1809265b5c not found: ID does not exist" Oct 02 07:45:37 crc kubenswrapper[4829]: I1002 07:45:37.475673 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" path="/var/lib/kubelet/pods/a796d106-1f58-418d-8dbb-5172dce0db93/volumes" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.078213 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkrcj"] Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.079440 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dkrcj" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="registry-server" containerID="cri-o://381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028" gracePeriod=2 Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.571347 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.680469 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-utilities\") pod \"1e082a97-b508-49a7-86d2-57c2453dd7b8\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.681365 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-catalog-content\") pod \"1e082a97-b508-49a7-86d2-57c2453dd7b8\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.681473 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v665m\" (UniqueName: \"kubernetes.io/projected/1e082a97-b508-49a7-86d2-57c2453dd7b8-kube-api-access-v665m\") pod \"1e082a97-b508-49a7-86d2-57c2453dd7b8\" (UID: \"1e082a97-b508-49a7-86d2-57c2453dd7b8\") " Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.681829 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-utilities" (OuterVolumeSpecName: "utilities") pod "1e082a97-b508-49a7-86d2-57c2453dd7b8" (UID: "1e082a97-b508-49a7-86d2-57c2453dd7b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.682541 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.692765 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e082a97-b508-49a7-86d2-57c2453dd7b8-kube-api-access-v665m" (OuterVolumeSpecName: "kube-api-access-v665m") pod "1e082a97-b508-49a7-86d2-57c2453dd7b8" (UID: "1e082a97-b508-49a7-86d2-57c2453dd7b8"). InnerVolumeSpecName "kube-api-access-v665m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.698365 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e082a97-b508-49a7-86d2-57c2453dd7b8" (UID: "1e082a97-b508-49a7-86d2-57c2453dd7b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.784161 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e082a97-b508-49a7-86d2-57c2453dd7b8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:39 crc kubenswrapper[4829]: I1002 07:45:39.784201 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v665m\" (UniqueName: \"kubernetes.io/projected/1e082a97-b508-49a7-86d2-57c2453dd7b8-kube-api-access-v665m\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.213193 4829 generic.go:334] "Generic (PLEG): container finished" podID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerID="381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028" exitCode=0 Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.213250 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkrcj" event={"ID":"1e082a97-b508-49a7-86d2-57c2453dd7b8","Type":"ContainerDied","Data":"381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028"} Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.213264 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dkrcj" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.213291 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dkrcj" event={"ID":"1e082a97-b508-49a7-86d2-57c2453dd7b8","Type":"ContainerDied","Data":"7678186818a1c99f10daa350b209b4cd8c50fee726a8e1f58cc2c781e641ac4a"} Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.213313 4829 scope.go:117] "RemoveContainer" containerID="381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.245308 4829 scope.go:117] "RemoveContainer" containerID="5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.246025 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkrcj"] Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.253970 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dkrcj"] Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.280057 4829 scope.go:117] "RemoveContainer" containerID="1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.324167 4829 scope.go:117] "RemoveContainer" containerID="381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028" Oct 02 07:45:40 crc kubenswrapper[4829]: E1002 07:45:40.324635 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028\": container with ID starting with 381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028 not found: ID does not exist" containerID="381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.324742 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028"} err="failed to get container status \"381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028\": rpc error: code = NotFound desc = could not find container \"381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028\": container with ID starting with 381ed1e23e242081612286a18168373a3d63f4af463b52a92b0d9042c3cc5028 not found: ID does not exist" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.324826 4829 scope.go:117] "RemoveContainer" containerID="5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8" Oct 02 07:45:40 crc kubenswrapper[4829]: E1002 07:45:40.325315 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8\": container with ID starting with 5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8 not found: ID does not exist" containerID="5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.325345 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8"} err="failed to get container status \"5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8\": rpc error: code = NotFound desc = could not find container \"5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8\": container with ID starting with 5b1112f31b55374373626516439e083de3c1c98e471d6cf91f206317d83c54b8 not found: ID does not exist" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.325366 4829 scope.go:117] "RemoveContainer" containerID="1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832" Oct 02 07:45:40 crc kubenswrapper[4829]: E1002 07:45:40.325644 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832\": container with ID starting with 1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832 not found: ID does not exist" containerID="1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832" Oct 02 07:45:40 crc kubenswrapper[4829]: I1002 07:45:40.325688 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832"} err="failed to get container status \"1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832\": rpc error: code = NotFound desc = could not find container \"1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832\": container with ID starting with 1b04a4dc3f7e5671a2b7333d6e8e3d1053d89f8fdb6ee543abfed95a7dc5e832 not found: ID does not exist" Oct 02 07:45:41 crc kubenswrapper[4829]: I1002 07:45:41.293110 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:41 crc kubenswrapper[4829]: I1002 07:45:41.384968 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:41 crc kubenswrapper[4829]: I1002 07:45:41.474690 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" path="/var/lib/kubelet/pods/1e082a97-b508-49a7-86d2-57c2453dd7b8/volumes" Oct 02 07:45:44 crc kubenswrapper[4829]: I1002 07:45:44.683696 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hv2nx"] Oct 02 07:45:44 crc kubenswrapper[4829]: I1002 07:45:44.684865 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hv2nx" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="registry-server" containerID="cri-o://13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9" gracePeriod=2 Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.174926 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.273503 4829 generic.go:334] "Generic (PLEG): container finished" podID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerID="13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9" exitCode=0 Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.273563 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hv2nx" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.273597 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerDied","Data":"13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9"} Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.273655 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hv2nx" event={"ID":"af87ca08-b9dd-4363-ab52-3f9605c309c2","Type":"ContainerDied","Data":"1b8e6b5e1280a49b7cd2c6468a01c0eaefd89be574c5882320faee5bd139aefe"} Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.273687 4829 scope.go:117] "RemoveContainer" containerID="13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.292502 4829 scope.go:117] "RemoveContainer" containerID="09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.307382 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-php8n\" (UniqueName: \"kubernetes.io/projected/af87ca08-b9dd-4363-ab52-3f9605c309c2-kube-api-access-php8n\") pod \"af87ca08-b9dd-4363-ab52-3f9605c309c2\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.307651 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-utilities\") pod \"af87ca08-b9dd-4363-ab52-3f9605c309c2\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.307741 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-catalog-content\") pod \"af87ca08-b9dd-4363-ab52-3f9605c309c2\" (UID: \"af87ca08-b9dd-4363-ab52-3f9605c309c2\") " Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.308427 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-utilities" (OuterVolumeSpecName: "utilities") pod "af87ca08-b9dd-4363-ab52-3f9605c309c2" (UID: "af87ca08-b9dd-4363-ab52-3f9605c309c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.313558 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af87ca08-b9dd-4363-ab52-3f9605c309c2-kube-api-access-php8n" (OuterVolumeSpecName: "kube-api-access-php8n") pod "af87ca08-b9dd-4363-ab52-3f9605c309c2" (UID: "af87ca08-b9dd-4363-ab52-3f9605c309c2"). InnerVolumeSpecName "kube-api-access-php8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.327241 4829 scope.go:117] "RemoveContainer" containerID="bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.357884 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af87ca08-b9dd-4363-ab52-3f9605c309c2" (UID: "af87ca08-b9dd-4363-ab52-3f9605c309c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.410203 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.410246 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-php8n\" (UniqueName: \"kubernetes.io/projected/af87ca08-b9dd-4363-ab52-3f9605c309c2-kube-api-access-php8n\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.410258 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af87ca08-b9dd-4363-ab52-3f9605c309c2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.442630 4829 scope.go:117] "RemoveContainer" containerID="13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9" Oct 02 07:45:45 crc kubenswrapper[4829]: E1002 07:45:45.443221 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9\": container with ID starting with 13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9 not found: ID does not exist" containerID="13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.443285 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9"} err="failed to get container status \"13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9\": rpc error: code = NotFound desc = could not find container \"13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9\": container with ID starting with 13038e0924c1649dc14f4df3beb0b56948044d4525f7cfb74e7e3e33b9015bc9 not found: ID does not exist" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.443317 4829 scope.go:117] "RemoveContainer" containerID="09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd" Oct 02 07:45:45 crc kubenswrapper[4829]: E1002 07:45:45.443852 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd\": container with ID starting with 09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd not found: ID does not exist" containerID="09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.443890 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd"} err="failed to get container status \"09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd\": rpc error: code = NotFound desc = could not find container \"09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd\": container with ID starting with 09eca09a2c9c70350ec27c143289bf972ea159eedcebb09c10cda674c26a25cd not found: ID does not exist" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.443925 4829 scope.go:117] "RemoveContainer" containerID="bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317" Oct 02 07:45:45 crc kubenswrapper[4829]: E1002 07:45:45.444730 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317\": container with ID starting with bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317 not found: ID does not exist" containerID="bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.444761 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317"} err="failed to get container status \"bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317\": rpc error: code = NotFound desc = could not find container \"bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317\": container with ID starting with bb2cf9c60bf834a66c50a0b2b48eaf0e8327d603a78082248e37787effc97317 not found: ID does not exist" Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.598043 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hv2nx"] Oct 02 07:45:45 crc kubenswrapper[4829]: I1002 07:45:45.605071 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hv2nx"] Oct 02 07:45:46 crc kubenswrapper[4829]: I1002 07:45:46.462121 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:45:46 crc kubenswrapper[4829]: E1002 07:45:46.463130 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:45:47 crc kubenswrapper[4829]: I1002 07:45:47.473837 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" path="/var/lib/kubelet/pods/af87ca08-b9dd-4363-ab52-3f9605c309c2/volumes" Oct 02 07:45:51 crc kubenswrapper[4829]: I1002 07:45:51.043573 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-w9kh7"] Oct 02 07:45:51 crc kubenswrapper[4829]: I1002 07:45:51.055736 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-d754t"] Oct 02 07:45:51 crc kubenswrapper[4829]: I1002 07:45:51.063060 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-w9kh7"] Oct 02 07:45:51 crc kubenswrapper[4829]: I1002 07:45:51.069875 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-d754t"] Oct 02 07:45:51 crc kubenswrapper[4829]: I1002 07:45:51.470863 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4509f44c-18ac-449a-9b64-619c7c304e16" path="/var/lib/kubelet/pods/4509f44c-18ac-449a-9b64-619c7c304e16/volumes" Oct 02 07:45:51 crc kubenswrapper[4829]: I1002 07:45:51.471483 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59c48271-56c9-483d-8db7-0e9cb785e2d2" path="/var/lib/kubelet/pods/59c48271-56c9-483d-8db7-0e9cb785e2d2/volumes" Oct 02 07:45:52 crc kubenswrapper[4829]: I1002 07:45:52.029740 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-sqbkw"] Oct 02 07:45:52 crc kubenswrapper[4829]: I1002 07:45:52.040195 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-sqbkw"] Oct 02 07:45:53 crc kubenswrapper[4829]: I1002 07:45:53.490323 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5be0189-8767-4668-b280-7cefa2f9237d" path="/var/lib/kubelet/pods/c5be0189-8767-4668-b280-7cefa2f9237d/volumes" Oct 02 07:45:57 crc kubenswrapper[4829]: I1002 07:45:57.857675 4829 scope.go:117] "RemoveContainer" containerID="da10059a3c91c82cf054a9566d982591ce0f1d1cd01453a06405b8c20b831c66" Oct 02 07:45:57 crc kubenswrapper[4829]: I1002 07:45:57.928380 4829 scope.go:117] "RemoveContainer" containerID="144cc00dadc81fb60e62fb7d52c52e4cacad8f37f9d76931e3e67f9f408d84a3" Oct 02 07:45:57 crc kubenswrapper[4829]: I1002 07:45:57.960405 4829 scope.go:117] "RemoveContainer" containerID="063a9351068734af6e6fa6977f5cd04774af913fd826b3020b82866f57834828" Oct 02 07:45:58 crc kubenswrapper[4829]: I1002 07:45:58.039003 4829 scope.go:117] "RemoveContainer" containerID="c9e91ac655fc2f9771cef316a9d9401d7568fc0fa938f60f8c095dc36caec018" Oct 02 07:45:58 crc kubenswrapper[4829]: I1002 07:45:58.078180 4829 scope.go:117] "RemoveContainer" containerID="9a91941185e1b0c50fbbe0c382ab6f2c4f9da63274f907f001279591b7beb02c" Oct 02 07:45:58 crc kubenswrapper[4829]: I1002 07:45:58.131247 4829 scope.go:117] "RemoveContainer" containerID="1d7e5fded6980d5fdaa53714f64a89ae74e49847ed59fcecd6a4f4a60686d2fa" Oct 02 07:45:58 crc kubenswrapper[4829]: I1002 07:45:58.179963 4829 scope.go:117] "RemoveContainer" containerID="1d3af3a6768c0582737fd2ee09f23f7028a593a6df44b6f0afe97a463949f90c" Oct 02 07:45:59 crc kubenswrapper[4829]: I1002 07:45:59.474407 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:45:59 crc kubenswrapper[4829]: E1002 07:45:59.474974 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:46:02 crc kubenswrapper[4829]: I1002 07:46:02.453671 4829 generic.go:334] "Generic (PLEG): container finished" podID="e6b2c2b0-9a68-4525-9f18-64b8e4354246" containerID="e17ea8a28fc9d46d18e840c410b53e2e0d63455cd065bbed012353451e925a88" exitCode=0 Oct 02 07:46:02 crc kubenswrapper[4829]: I1002 07:46:02.453822 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" event={"ID":"e6b2c2b0-9a68-4525-9f18-64b8e4354246","Type":"ContainerDied","Data":"e17ea8a28fc9d46d18e840c410b53e2e0d63455cd065bbed012353451e925a88"} Oct 02 07:46:03 crc kubenswrapper[4829]: I1002 07:46:03.871642 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.042515 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75xcv\" (UniqueName: \"kubernetes.io/projected/e6b2c2b0-9a68-4525-9f18-64b8e4354246-kube-api-access-75xcv\") pod \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.042619 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-ssh-key\") pod \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.042788 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-inventory\") pod \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\" (UID: \"e6b2c2b0-9a68-4525-9f18-64b8e4354246\") " Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.055058 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6b2c2b0-9a68-4525-9f18-64b8e4354246-kube-api-access-75xcv" (OuterVolumeSpecName: "kube-api-access-75xcv") pod "e6b2c2b0-9a68-4525-9f18-64b8e4354246" (UID: "e6b2c2b0-9a68-4525-9f18-64b8e4354246"). InnerVolumeSpecName "kube-api-access-75xcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.091596 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6b2c2b0-9a68-4525-9f18-64b8e4354246" (UID: "e6b2c2b0-9a68-4525-9f18-64b8e4354246"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.093928 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-inventory" (OuterVolumeSpecName: "inventory") pod "e6b2c2b0-9a68-4525-9f18-64b8e4354246" (UID: "e6b2c2b0-9a68-4525-9f18-64b8e4354246"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.145949 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75xcv\" (UniqueName: \"kubernetes.io/projected/e6b2c2b0-9a68-4525-9f18-64b8e4354246-kube-api-access-75xcv\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.145995 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.146011 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b2c2b0-9a68-4525-9f18-64b8e4354246-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.478772 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" event={"ID":"e6b2c2b0-9a68-4525-9f18-64b8e4354246","Type":"ContainerDied","Data":"9003241f7ce36c51fb11a88827f55aa1d38ecb1ba0c732c2d9173209214a192c"} Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.478872 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9003241f7ce36c51fb11a88827f55aa1d38ecb1ba0c732c2d9173209214a192c" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.478823 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.602385 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5"] Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.603103 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="extract-utilities" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.603206 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="extract-utilities" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.603309 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.603401 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.603497 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="extract-content" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.603572 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="extract-content" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.603652 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.603720 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.603797 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="extract-utilities" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.603876 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="extract-utilities" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.603971 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="extract-content" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604042 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="extract-content" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.604124 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604197 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.604307 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6b2c2b0-9a68-4525-9f18-64b8e4354246" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604379 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6b2c2b0-9a68-4525-9f18-64b8e4354246" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.604463 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="extract-content" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604531 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="extract-content" Oct 02 07:46:04 crc kubenswrapper[4829]: E1002 07:46:04.604588 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="extract-utilities" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604645 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="extract-utilities" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604902 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6b2c2b0-9a68-4525-9f18-64b8e4354246" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.604978 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e082a97-b508-49a7-86d2-57c2453dd7b8" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.605045 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="af87ca08-b9dd-4363-ab52-3f9605c309c2" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.605114 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="a796d106-1f58-418d-8dbb-5172dce0db93" containerName="registry-server" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.605816 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.608507 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.611134 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.611494 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.611778 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.615077 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5"] Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.759515 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cqfh\" (UniqueName: \"kubernetes.io/projected/bb87d295-2bab-44ea-b5cc-cc77a0efe964-kube-api-access-9cqfh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.759701 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.759788 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.861936 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.862051 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.862283 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cqfh\" (UniqueName: \"kubernetes.io/projected/bb87d295-2bab-44ea-b5cc-cc77a0efe964-kube-api-access-9cqfh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.866764 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.875677 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.891393 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cqfh\" (UniqueName: \"kubernetes.io/projected/bb87d295-2bab-44ea-b5cc-cc77a0efe964-kube-api-access-9cqfh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:04 crc kubenswrapper[4829]: I1002 07:46:04.942449 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:05 crc kubenswrapper[4829]: I1002 07:46:05.586088 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5"] Oct 02 07:46:06 crc kubenswrapper[4829]: I1002 07:46:06.517401 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" event={"ID":"bb87d295-2bab-44ea-b5cc-cc77a0efe964","Type":"ContainerStarted","Data":"6df8d0baf947e8dcd4ff31fa9458f41ee6426606085d2ab12f570724afdf0a1e"} Oct 02 07:46:06 crc kubenswrapper[4829]: I1002 07:46:06.517726 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" event={"ID":"bb87d295-2bab-44ea-b5cc-cc77a0efe964","Type":"ContainerStarted","Data":"c1c9fdbfac21f674de093a0db2a7cf8700514755c743a8e3fdf75bf33d9d9018"} Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.038864 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" podStartSLOduration=4.5833990159999995 podStartE2EDuration="5.038842829s" podCreationTimestamp="2025-10-02 07:46:04 +0000 UTC" firstStartedPulling="2025-10-02 07:46:05.603761276 +0000 UTC m=+1756.943409691" lastFinishedPulling="2025-10-02 07:46:06.059205069 +0000 UTC m=+1757.398853504" observedRunningTime="2025-10-02 07:46:06.547778613 +0000 UTC m=+1757.887427048" watchObservedRunningTime="2025-10-02 07:46:09.038842829 +0000 UTC m=+1760.378491234" Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.041626 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-3b7f-account-create-dpq82"] Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.051428 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-faf0-account-create-qxxsr"] Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.059649 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0c73-account-create-648bq"] Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.067496 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-3b7f-account-create-dpq82"] Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.077524 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-faf0-account-create-qxxsr"] Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.086371 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0c73-account-create-648bq"] Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.473762 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19037609-49d9-4e96-b537-30e2765c6f26" path="/var/lib/kubelet/pods/19037609-49d9-4e96-b537-30e2765c6f26/volumes" Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.474314 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9460237e-d366-4526-a859-06daeae67fce" path="/var/lib/kubelet/pods/9460237e-d366-4526-a859-06daeae67fce/volumes" Oct 02 07:46:09 crc kubenswrapper[4829]: I1002 07:46:09.474766 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecce1ed1-3a3e-4576-853e-0972201544e9" path="/var/lib/kubelet/pods/ecce1ed1-3a3e-4576-853e-0972201544e9/volumes" Oct 02 07:46:12 crc kubenswrapper[4829]: I1002 07:46:12.593066 4829 generic.go:334] "Generic (PLEG): container finished" podID="bb87d295-2bab-44ea-b5cc-cc77a0efe964" containerID="6df8d0baf947e8dcd4ff31fa9458f41ee6426606085d2ab12f570724afdf0a1e" exitCode=0 Oct 02 07:46:12 crc kubenswrapper[4829]: I1002 07:46:12.593185 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" event={"ID":"bb87d295-2bab-44ea-b5cc-cc77a0efe964","Type":"ContainerDied","Data":"6df8d0baf947e8dcd4ff31fa9458f41ee6426606085d2ab12f570724afdf0a1e"} Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.102737 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.155295 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-inventory\") pod \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.155367 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-ssh-key\") pod \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.155500 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cqfh\" (UniqueName: \"kubernetes.io/projected/bb87d295-2bab-44ea-b5cc-cc77a0efe964-kube-api-access-9cqfh\") pod \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\" (UID: \"bb87d295-2bab-44ea-b5cc-cc77a0efe964\") " Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.166429 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb87d295-2bab-44ea-b5cc-cc77a0efe964-kube-api-access-9cqfh" (OuterVolumeSpecName: "kube-api-access-9cqfh") pod "bb87d295-2bab-44ea-b5cc-cc77a0efe964" (UID: "bb87d295-2bab-44ea-b5cc-cc77a0efe964"). InnerVolumeSpecName "kube-api-access-9cqfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.188952 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-inventory" (OuterVolumeSpecName: "inventory") pod "bb87d295-2bab-44ea-b5cc-cc77a0efe964" (UID: "bb87d295-2bab-44ea-b5cc-cc77a0efe964"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.191756 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bb87d295-2bab-44ea-b5cc-cc77a0efe964" (UID: "bb87d295-2bab-44ea-b5cc-cc77a0efe964"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.258031 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.258063 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb87d295-2bab-44ea-b5cc-cc77a0efe964-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.258074 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cqfh\" (UniqueName: \"kubernetes.io/projected/bb87d295-2bab-44ea-b5cc-cc77a0efe964-kube-api-access-9cqfh\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.618307 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" event={"ID":"bb87d295-2bab-44ea-b5cc-cc77a0efe964","Type":"ContainerDied","Data":"c1c9fdbfac21f674de093a0db2a7cf8700514755c743a8e3fdf75bf33d9d9018"} Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.618523 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1c9fdbfac21f674de093a0db2a7cf8700514755c743a8e3fdf75bf33d9d9018" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.618431 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.702593 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw"] Oct 02 07:46:14 crc kubenswrapper[4829]: E1002 07:46:14.703330 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb87d295-2bab-44ea-b5cc-cc77a0efe964" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.703426 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb87d295-2bab-44ea-b5cc-cc77a0efe964" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.703774 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb87d295-2bab-44ea-b5cc-cc77a0efe964" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.704641 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.709205 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.709400 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.709487 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.709460 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.736642 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw"] Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.768498 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.768671 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs548\" (UniqueName: \"kubernetes.io/projected/4181a86c-75a3-4400-9b89-352c04c06044-kube-api-access-vs548\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.768715 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.870697 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs548\" (UniqueName: \"kubernetes.io/projected/4181a86c-75a3-4400-9b89-352c04c06044-kube-api-access-vs548\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.870793 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.870928 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.876800 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.884992 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:14 crc kubenswrapper[4829]: I1002 07:46:14.890155 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs548\" (UniqueName: \"kubernetes.io/projected/4181a86c-75a3-4400-9b89-352c04c06044-kube-api-access-vs548\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-dfmsw\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:15 crc kubenswrapper[4829]: I1002 07:46:15.045136 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:15 crc kubenswrapper[4829]: I1002 07:46:15.461219 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:46:15 crc kubenswrapper[4829]: E1002 07:46:15.461858 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:46:15 crc kubenswrapper[4829]: I1002 07:46:15.673089 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw"] Oct 02 07:46:15 crc kubenswrapper[4829]: W1002 07:46:15.679419 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4181a86c_75a3_4400_9b89_352c04c06044.slice/crio-b991e48ab3bc52f99ffc6eaf0ecf92f2e86507b4c6cc5e77a0c8f8bb649d02d1 WatchSource:0}: Error finding container b991e48ab3bc52f99ffc6eaf0ecf92f2e86507b4c6cc5e77a0c8f8bb649d02d1: Status 404 returned error can't find the container with id b991e48ab3bc52f99ffc6eaf0ecf92f2e86507b4c6cc5e77a0c8f8bb649d02d1 Oct 02 07:46:16 crc kubenswrapper[4829]: I1002 07:46:16.641962 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" event={"ID":"4181a86c-75a3-4400-9b89-352c04c06044","Type":"ContainerStarted","Data":"c4f2807ab458528a353ac4e46173c14ce7f7efc5abc2235cba62f16e30f424e2"} Oct 02 07:46:16 crc kubenswrapper[4829]: I1002 07:46:16.642391 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" event={"ID":"4181a86c-75a3-4400-9b89-352c04c06044","Type":"ContainerStarted","Data":"b991e48ab3bc52f99ffc6eaf0ecf92f2e86507b4c6cc5e77a0c8f8bb649d02d1"} Oct 02 07:46:16 crc kubenswrapper[4829]: I1002 07:46:16.670625 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" podStartSLOduration=2.173654536 podStartE2EDuration="2.670604201s" podCreationTimestamp="2025-10-02 07:46:14 +0000 UTC" firstStartedPulling="2025-10-02 07:46:15.681271962 +0000 UTC m=+1767.020920367" lastFinishedPulling="2025-10-02 07:46:16.178221627 +0000 UTC m=+1767.517870032" observedRunningTime="2025-10-02 07:46:16.666708129 +0000 UTC m=+1768.006356544" watchObservedRunningTime="2025-10-02 07:46:16.670604201 +0000 UTC m=+1768.010252616" Oct 02 07:46:26 crc kubenswrapper[4829]: I1002 07:46:26.488833 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:46:26 crc kubenswrapper[4829]: E1002 07:46:26.489586 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:46:31 crc kubenswrapper[4829]: I1002 07:46:31.032046 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8fgfl"] Oct 02 07:46:31 crc kubenswrapper[4829]: I1002 07:46:31.039697 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-8fgfl"] Oct 02 07:46:31 crc kubenswrapper[4829]: I1002 07:46:31.475471 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f49fa10e-56bf-4af5-874a-9e73967cd27d" path="/var/lib/kubelet/pods/f49fa10e-56bf-4af5-874a-9e73967cd27d/volumes" Oct 02 07:46:39 crc kubenswrapper[4829]: I1002 07:46:39.467211 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:46:39 crc kubenswrapper[4829]: E1002 07:46:39.469056 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:46:51 crc kubenswrapper[4829]: I1002 07:46:51.461654 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:46:51 crc kubenswrapper[4829]: E1002 07:46:51.462696 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:46:53 crc kubenswrapper[4829]: I1002 07:46:53.080836 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-4jqqc"] Oct 02 07:46:53 crc kubenswrapper[4829]: I1002 07:46:53.093271 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-4jqqc"] Oct 02 07:46:53 crc kubenswrapper[4829]: I1002 07:46:53.472035 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f195e99-4316-4030-9e49-1bf368defe41" path="/var/lib/kubelet/pods/0f195e99-4316-4030-9e49-1bf368defe41/volumes" Oct 02 07:46:54 crc kubenswrapper[4829]: I1002 07:46:54.040128 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlwp8"] Oct 02 07:46:54 crc kubenswrapper[4829]: I1002 07:46:54.054676 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-zlwp8"] Oct 02 07:46:55 crc kubenswrapper[4829]: I1002 07:46:55.472900 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d582f285-53ae-41fd-9be1-9b18dd9e95f9" path="/var/lib/kubelet/pods/d582f285-53ae-41fd-9be1-9b18dd9e95f9/volumes" Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.076645 4829 generic.go:334] "Generic (PLEG): container finished" podID="4181a86c-75a3-4400-9b89-352c04c06044" containerID="c4f2807ab458528a353ac4e46173c14ce7f7efc5abc2235cba62f16e30f424e2" exitCode=0 Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.076764 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" event={"ID":"4181a86c-75a3-4400-9b89-352c04c06044","Type":"ContainerDied","Data":"c4f2807ab458528a353ac4e46173c14ce7f7efc5abc2235cba62f16e30f424e2"} Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.405473 4829 scope.go:117] "RemoveContainer" containerID="4b2aba6614d0ef4b94061fb92b2bd700c14bbc24b3ab2c4572f13109875b58fd" Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.435843 4829 scope.go:117] "RemoveContainer" containerID="160849ffc45791c87a99b94a6fbd5c188e073902530fd11a5276ddd64b119939" Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.517565 4829 scope.go:117] "RemoveContainer" containerID="ba9fb13ff51c68b72c7b7787afa0b55465ee90a6af58d91e5626e2287e81a1ed" Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.578555 4829 scope.go:117] "RemoveContainer" containerID="fa3755afd0985e5c01dad6075ff0333806be74baa8948041accbb147d52e0b6c" Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.603673 4829 scope.go:117] "RemoveContainer" containerID="c1d874ae94d8d9d7f79e67f5c372a59c2c1c1b604d5709981151cef9e88a3428" Oct 02 07:46:58 crc kubenswrapper[4829]: I1002 07:46:58.645345 4829 scope.go:117] "RemoveContainer" containerID="787dc83cf5df9eefea49e19665d944403ada41509d7f27120d3529a358363a7b" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.581943 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.754032 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vs548\" (UniqueName: \"kubernetes.io/projected/4181a86c-75a3-4400-9b89-352c04c06044-kube-api-access-vs548\") pod \"4181a86c-75a3-4400-9b89-352c04c06044\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.754159 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-inventory\") pod \"4181a86c-75a3-4400-9b89-352c04c06044\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.757065 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-ssh-key\") pod \"4181a86c-75a3-4400-9b89-352c04c06044\" (UID: \"4181a86c-75a3-4400-9b89-352c04c06044\") " Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.761891 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4181a86c-75a3-4400-9b89-352c04c06044-kube-api-access-vs548" (OuterVolumeSpecName: "kube-api-access-vs548") pod "4181a86c-75a3-4400-9b89-352c04c06044" (UID: "4181a86c-75a3-4400-9b89-352c04c06044"). InnerVolumeSpecName "kube-api-access-vs548". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.786515 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4181a86c-75a3-4400-9b89-352c04c06044" (UID: "4181a86c-75a3-4400-9b89-352c04c06044"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.797519 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-inventory" (OuterVolumeSpecName: "inventory") pod "4181a86c-75a3-4400-9b89-352c04c06044" (UID: "4181a86c-75a3-4400-9b89-352c04c06044"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.860894 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.860933 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vs548\" (UniqueName: \"kubernetes.io/projected/4181a86c-75a3-4400-9b89-352c04c06044-kube-api-access-vs548\") on node \"crc\" DevicePath \"\"" Oct 02 07:46:59 crc kubenswrapper[4829]: I1002 07:46:59.860948 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4181a86c-75a3-4400-9b89-352c04c06044-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.098855 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" event={"ID":"4181a86c-75a3-4400-9b89-352c04c06044","Type":"ContainerDied","Data":"b991e48ab3bc52f99ffc6eaf0ecf92f2e86507b4c6cc5e77a0c8f8bb649d02d1"} Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.099202 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b991e48ab3bc52f99ffc6eaf0ecf92f2e86507b4c6cc5e77a0c8f8bb649d02d1" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.098901 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-dfmsw" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.188345 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m"] Oct 02 07:47:00 crc kubenswrapper[4829]: E1002 07:47:00.188843 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4181a86c-75a3-4400-9b89-352c04c06044" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.188867 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4181a86c-75a3-4400-9b89-352c04c06044" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.189093 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4181a86c-75a3-4400-9b89-352c04c06044" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.189815 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.191867 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.191913 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.193448 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.197956 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m"] Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.198184 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.269962 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.270312 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7557\" (UniqueName: \"kubernetes.io/projected/28b94275-3752-42ab-b862-fee2ee0ead47-kube-api-access-j7557\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.270473 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.372340 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.372418 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7557\" (UniqueName: \"kubernetes.io/projected/28b94275-3752-42ab-b862-fee2ee0ead47-kube-api-access-j7557\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.372472 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.377128 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.377885 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.393513 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7557\" (UniqueName: \"kubernetes.io/projected/28b94275-3752-42ab-b862-fee2ee0ead47-kube-api-access-j7557\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k746m\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:00 crc kubenswrapper[4829]: I1002 07:47:00.569842 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:47:01 crc kubenswrapper[4829]: I1002 07:47:01.141288 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m"] Oct 02 07:47:02 crc kubenswrapper[4829]: I1002 07:47:02.119648 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" event={"ID":"28b94275-3752-42ab-b862-fee2ee0ead47","Type":"ContainerStarted","Data":"2c2219f5fc411bb508de077efeeca6bd1ef08c90265d316a889e2078ebce46ca"} Oct 02 07:47:02 crc kubenswrapper[4829]: I1002 07:47:02.120107 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" event={"ID":"28b94275-3752-42ab-b862-fee2ee0ead47","Type":"ContainerStarted","Data":"0503806e6eb3fb60aedd349a9a3534cc33fdb1c8b45db5048fe0f81884104eba"} Oct 02 07:47:02 crc kubenswrapper[4829]: I1002 07:47:02.143705 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" podStartSLOduration=1.6378345570000001 podStartE2EDuration="2.143680219s" podCreationTimestamp="2025-10-02 07:47:00 +0000 UTC" firstStartedPulling="2025-10-02 07:47:01.144026239 +0000 UTC m=+1812.483674664" lastFinishedPulling="2025-10-02 07:47:01.649871901 +0000 UTC m=+1812.989520326" observedRunningTime="2025-10-02 07:47:02.137556868 +0000 UTC m=+1813.477205293" watchObservedRunningTime="2025-10-02 07:47:02.143680219 +0000 UTC m=+1813.483328664" Oct 02 07:47:06 crc kubenswrapper[4829]: I1002 07:47:06.461421 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:47:07 crc kubenswrapper[4829]: I1002 07:47:07.174624 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"46cd74959e46dfb078660ee3089447f9db6cde42ac4b590e00c9b5fae637513e"} Oct 02 07:47:39 crc kubenswrapper[4829]: I1002 07:47:39.061159 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-8drj4"] Oct 02 07:47:39 crc kubenswrapper[4829]: I1002 07:47:39.072866 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-8drj4"] Oct 02 07:47:39 crc kubenswrapper[4829]: I1002 07:47:39.485897 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d864f84e-5ba0-4fca-96e4-1fe63880960d" path="/var/lib/kubelet/pods/d864f84e-5ba0-4fca-96e4-1fe63880960d/volumes" Oct 02 07:47:58 crc kubenswrapper[4829]: I1002 07:47:58.861733 4829 scope.go:117] "RemoveContainer" containerID="6f98e26526394978d637124d05c083aa1dc0def656b4304adb9882cd3bd1afc3" Oct 02 07:47:59 crc kubenswrapper[4829]: I1002 07:47:59.765927 4829 generic.go:334] "Generic (PLEG): container finished" podID="28b94275-3752-42ab-b862-fee2ee0ead47" containerID="2c2219f5fc411bb508de077efeeca6bd1ef08c90265d316a889e2078ebce46ca" exitCode=2 Oct 02 07:47:59 crc kubenswrapper[4829]: I1002 07:47:59.766055 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" event={"ID":"28b94275-3752-42ab-b862-fee2ee0ead47","Type":"ContainerDied","Data":"2c2219f5fc411bb508de077efeeca6bd1ef08c90265d316a889e2078ebce46ca"} Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.192965 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.268691 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-inventory\") pod \"28b94275-3752-42ab-b862-fee2ee0ead47\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.269071 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-ssh-key\") pod \"28b94275-3752-42ab-b862-fee2ee0ead47\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.269331 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7557\" (UniqueName: \"kubernetes.io/projected/28b94275-3752-42ab-b862-fee2ee0ead47-kube-api-access-j7557\") pod \"28b94275-3752-42ab-b862-fee2ee0ead47\" (UID: \"28b94275-3752-42ab-b862-fee2ee0ead47\") " Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.274801 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28b94275-3752-42ab-b862-fee2ee0ead47-kube-api-access-j7557" (OuterVolumeSpecName: "kube-api-access-j7557") pod "28b94275-3752-42ab-b862-fee2ee0ead47" (UID: "28b94275-3752-42ab-b862-fee2ee0ead47"). InnerVolumeSpecName "kube-api-access-j7557". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.304274 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "28b94275-3752-42ab-b862-fee2ee0ead47" (UID: "28b94275-3752-42ab-b862-fee2ee0ead47"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.316985 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-inventory" (OuterVolumeSpecName: "inventory") pod "28b94275-3752-42ab-b862-fee2ee0ead47" (UID: "28b94275-3752-42ab-b862-fee2ee0ead47"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.372246 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.372470 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28b94275-3752-42ab-b862-fee2ee0ead47-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.372623 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7557\" (UniqueName: \"kubernetes.io/projected/28b94275-3752-42ab-b862-fee2ee0ead47-kube-api-access-j7557\") on node \"crc\" DevicePath \"\"" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.788843 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" event={"ID":"28b94275-3752-42ab-b862-fee2ee0ead47","Type":"ContainerDied","Data":"0503806e6eb3fb60aedd349a9a3534cc33fdb1c8b45db5048fe0f81884104eba"} Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.788895 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0503806e6eb3fb60aedd349a9a3534cc33fdb1c8b45db5048fe0f81884104eba" Oct 02 07:48:01 crc kubenswrapper[4829]: I1002 07:48:01.788950 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k746m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.038586 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m"] Oct 02 07:48:09 crc kubenswrapper[4829]: E1002 07:48:09.040376 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28b94275-3752-42ab-b862-fee2ee0ead47" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.040397 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="28b94275-3752-42ab-b862-fee2ee0ead47" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.040675 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="28b94275-3752-42ab-b862-fee2ee0ead47" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.041692 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.046756 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m"] Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.047317 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.047342 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.047443 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.048003 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.140751 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.140874 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.141135 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pns5r\" (UniqueName: \"kubernetes.io/projected/f418c44c-86bc-4ba6-aabb-da19658c0e77-kube-api-access-pns5r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.242881 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pns5r\" (UniqueName: \"kubernetes.io/projected/f418c44c-86bc-4ba6-aabb-da19658c0e77-kube-api-access-pns5r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.242980 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.243046 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.248790 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.255863 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.259752 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pns5r\" (UniqueName: \"kubernetes.io/projected/f418c44c-86bc-4ba6-aabb-da19658c0e77-kube-api-access-pns5r\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.397375 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.948163 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m"] Oct 02 07:48:09 crc kubenswrapper[4829]: I1002 07:48:09.957796 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:48:10 crc kubenswrapper[4829]: I1002 07:48:10.877178 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" event={"ID":"f418c44c-86bc-4ba6-aabb-da19658c0e77","Type":"ContainerStarted","Data":"5817fc94a47262370e37ac04de9660d6c564a1b3648154426b51a544e451666b"} Oct 02 07:48:10 crc kubenswrapper[4829]: I1002 07:48:10.879312 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" event={"ID":"f418c44c-86bc-4ba6-aabb-da19658c0e77","Type":"ContainerStarted","Data":"4c7f8882050e8a2c4c786c3b5b8cc65fbe545840eeb12dd6c28271c9cdddd8bb"} Oct 02 07:48:10 crc kubenswrapper[4829]: I1002 07:48:10.902554 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" podStartSLOduration=1.429529511 podStartE2EDuration="1.90253492s" podCreationTimestamp="2025-10-02 07:48:09 +0000 UTC" firstStartedPulling="2025-10-02 07:48:09.957512323 +0000 UTC m=+1881.297160738" lastFinishedPulling="2025-10-02 07:48:10.430517752 +0000 UTC m=+1881.770166147" observedRunningTime="2025-10-02 07:48:10.900550858 +0000 UTC m=+1882.240199273" watchObservedRunningTime="2025-10-02 07:48:10.90253492 +0000 UTC m=+1882.242183325" Oct 02 07:49:02 crc kubenswrapper[4829]: I1002 07:49:02.400581 4829 generic.go:334] "Generic (PLEG): container finished" podID="f418c44c-86bc-4ba6-aabb-da19658c0e77" containerID="5817fc94a47262370e37ac04de9660d6c564a1b3648154426b51a544e451666b" exitCode=0 Oct 02 07:49:02 crc kubenswrapper[4829]: I1002 07:49:02.400687 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" event={"ID":"f418c44c-86bc-4ba6-aabb-da19658c0e77","Type":"ContainerDied","Data":"5817fc94a47262370e37ac04de9660d6c564a1b3648154426b51a544e451666b"} Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.828932 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.918756 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-inventory\") pod \"f418c44c-86bc-4ba6-aabb-da19658c0e77\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.918824 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-ssh-key\") pod \"f418c44c-86bc-4ba6-aabb-da19658c0e77\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.919111 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pns5r\" (UniqueName: \"kubernetes.io/projected/f418c44c-86bc-4ba6-aabb-da19658c0e77-kube-api-access-pns5r\") pod \"f418c44c-86bc-4ba6-aabb-da19658c0e77\" (UID: \"f418c44c-86bc-4ba6-aabb-da19658c0e77\") " Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.924420 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f418c44c-86bc-4ba6-aabb-da19658c0e77-kube-api-access-pns5r" (OuterVolumeSpecName: "kube-api-access-pns5r") pod "f418c44c-86bc-4ba6-aabb-da19658c0e77" (UID: "f418c44c-86bc-4ba6-aabb-da19658c0e77"). InnerVolumeSpecName "kube-api-access-pns5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.948615 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-inventory" (OuterVolumeSpecName: "inventory") pod "f418c44c-86bc-4ba6-aabb-da19658c0e77" (UID: "f418c44c-86bc-4ba6-aabb-da19658c0e77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:03 crc kubenswrapper[4829]: I1002 07:49:03.948954 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f418c44c-86bc-4ba6-aabb-da19658c0e77" (UID: "f418c44c-86bc-4ba6-aabb-da19658c0e77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.022117 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.022476 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f418c44c-86bc-4ba6-aabb-da19658c0e77-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.022496 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pns5r\" (UniqueName: \"kubernetes.io/projected/f418c44c-86bc-4ba6-aabb-da19658c0e77-kube-api-access-pns5r\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.423891 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" event={"ID":"f418c44c-86bc-4ba6-aabb-da19658c0e77","Type":"ContainerDied","Data":"4c7f8882050e8a2c4c786c3b5b8cc65fbe545840eeb12dd6c28271c9cdddd8bb"} Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.423943 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c7f8882050e8a2c4c786c3b5b8cc65fbe545840eeb12dd6c28271c9cdddd8bb" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.423988 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.543976 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8b6wr"] Oct 02 07:49:04 crc kubenswrapper[4829]: E1002 07:49:04.544582 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f418c44c-86bc-4ba6-aabb-da19658c0e77" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.544608 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="f418c44c-86bc-4ba6-aabb-da19658c0e77" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.545687 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="f418c44c-86bc-4ba6-aabb-da19658c0e77" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.546586 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.556384 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.557191 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.557249 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.557200 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.558105 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8b6wr"] Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.636932 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvfbx\" (UniqueName: \"kubernetes.io/projected/0506e251-5b72-4ce4-91c5-8116bf387c34-kube-api-access-cvfbx\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.637014 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.637079 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.739663 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvfbx\" (UniqueName: \"kubernetes.io/projected/0506e251-5b72-4ce4-91c5-8116bf387c34-kube-api-access-cvfbx\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.739812 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.739842 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.745959 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.746012 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.759825 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvfbx\" (UniqueName: \"kubernetes.io/projected/0506e251-5b72-4ce4-91c5-8116bf387c34-kube-api-access-cvfbx\") pod \"ssh-known-hosts-edpm-deployment-8b6wr\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:04 crc kubenswrapper[4829]: I1002 07:49:04.873944 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:05 crc kubenswrapper[4829]: I1002 07:49:05.459598 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-8b6wr"] Oct 02 07:49:06 crc kubenswrapper[4829]: I1002 07:49:06.446149 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" event={"ID":"0506e251-5b72-4ce4-91c5-8116bf387c34","Type":"ContainerStarted","Data":"a257575e401e81020fb0ea5de7d323193887ea957ebd1b00d296c81c92f0862c"} Oct 02 07:49:06 crc kubenswrapper[4829]: I1002 07:49:06.446619 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" event={"ID":"0506e251-5b72-4ce4-91c5-8116bf387c34","Type":"ContainerStarted","Data":"f3e34135b8f8abfe696de08a97c2725f859bbd498e43b86fe30af3cbeec5ee53"} Oct 02 07:49:06 crc kubenswrapper[4829]: I1002 07:49:06.467042 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" podStartSLOduration=1.9863863400000001 podStartE2EDuration="2.467020887s" podCreationTimestamp="2025-10-02 07:49:04 +0000 UTC" firstStartedPulling="2025-10-02 07:49:05.462771842 +0000 UTC m=+1936.802420267" lastFinishedPulling="2025-10-02 07:49:05.943406409 +0000 UTC m=+1937.283054814" observedRunningTime="2025-10-02 07:49:06.463859118 +0000 UTC m=+1937.803507553" watchObservedRunningTime="2025-10-02 07:49:06.467020887 +0000 UTC m=+1937.806669292" Oct 02 07:49:14 crc kubenswrapper[4829]: I1002 07:49:14.530801 4829 generic.go:334] "Generic (PLEG): container finished" podID="0506e251-5b72-4ce4-91c5-8116bf387c34" containerID="a257575e401e81020fb0ea5de7d323193887ea957ebd1b00d296c81c92f0862c" exitCode=0 Oct 02 07:49:14 crc kubenswrapper[4829]: I1002 07:49:14.530919 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" event={"ID":"0506e251-5b72-4ce4-91c5-8116bf387c34","Type":"ContainerDied","Data":"a257575e401e81020fb0ea5de7d323193887ea957ebd1b00d296c81c92f0862c"} Oct 02 07:49:15 crc kubenswrapper[4829]: I1002 07:49:15.991463 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.074049 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvfbx\" (UniqueName: \"kubernetes.io/projected/0506e251-5b72-4ce4-91c5-8116bf387c34-kube-api-access-cvfbx\") pod \"0506e251-5b72-4ce4-91c5-8116bf387c34\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.074141 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-ssh-key-openstack-edpm-ipam\") pod \"0506e251-5b72-4ce4-91c5-8116bf387c34\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.074167 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-inventory-0\") pod \"0506e251-5b72-4ce4-91c5-8116bf387c34\" (UID: \"0506e251-5b72-4ce4-91c5-8116bf387c34\") " Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.083577 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0506e251-5b72-4ce4-91c5-8116bf387c34-kube-api-access-cvfbx" (OuterVolumeSpecName: "kube-api-access-cvfbx") pod "0506e251-5b72-4ce4-91c5-8116bf387c34" (UID: "0506e251-5b72-4ce4-91c5-8116bf387c34"). InnerVolumeSpecName "kube-api-access-cvfbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.117573 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "0506e251-5b72-4ce4-91c5-8116bf387c34" (UID: "0506e251-5b72-4ce4-91c5-8116bf387c34"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.141726 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0506e251-5b72-4ce4-91c5-8116bf387c34" (UID: "0506e251-5b72-4ce4-91c5-8116bf387c34"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.177633 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvfbx\" (UniqueName: \"kubernetes.io/projected/0506e251-5b72-4ce4-91c5-8116bf387c34-kube-api-access-cvfbx\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.177678 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.177691 4829 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/0506e251-5b72-4ce4-91c5-8116bf387c34-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.552528 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" event={"ID":"0506e251-5b72-4ce4-91c5-8116bf387c34","Type":"ContainerDied","Data":"f3e34135b8f8abfe696de08a97c2725f859bbd498e43b86fe30af3cbeec5ee53"} Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.552568 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3e34135b8f8abfe696de08a97c2725f859bbd498e43b86fe30af3cbeec5ee53" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.552583 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-8b6wr" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.666348 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj"] Oct 02 07:49:16 crc kubenswrapper[4829]: E1002 07:49:16.666822 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0506e251-5b72-4ce4-91c5-8116bf387c34" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.666847 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0506e251-5b72-4ce4-91c5-8116bf387c34" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.667110 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0506e251-5b72-4ce4-91c5-8116bf387c34" containerName="ssh-known-hosts-edpm-deployment" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.667803 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.669628 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.670069 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.675025 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.675731 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.683717 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj"] Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.789540 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx9hq\" (UniqueName: \"kubernetes.io/projected/3ed1e719-596e-42fd-ac3f-b092996c9ca6-kube-api-access-jx9hq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.790046 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.790072 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.892293 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.892338 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.892480 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx9hq\" (UniqueName: \"kubernetes.io/projected/3ed1e719-596e-42fd-ac3f-b092996c9ca6-kube-api-access-jx9hq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.898414 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.898534 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.912275 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx9hq\" (UniqueName: \"kubernetes.io/projected/3ed1e719-596e-42fd-ac3f-b092996c9ca6-kube-api-access-jx9hq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2kgpj\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:16 crc kubenswrapper[4829]: I1002 07:49:16.996488 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:17 crc kubenswrapper[4829]: I1002 07:49:17.540634 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj"] Oct 02 07:49:17 crc kubenswrapper[4829]: I1002 07:49:17.570732 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" event={"ID":"3ed1e719-596e-42fd-ac3f-b092996c9ca6","Type":"ContainerStarted","Data":"12888828ec842ee18ef1a13d268ec18277d6272b8ca1e479812f8a1d7b84c8f1"} Oct 02 07:49:18 crc kubenswrapper[4829]: I1002 07:49:18.582972 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" event={"ID":"3ed1e719-596e-42fd-ac3f-b092996c9ca6","Type":"ContainerStarted","Data":"520e3e77500d848b6eea90ac2b75abf86a0afa3e939bf03770ed299b5faeea72"} Oct 02 07:49:18 crc kubenswrapper[4829]: I1002 07:49:18.613758 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" podStartSLOduration=2.097521783 podStartE2EDuration="2.613731128s" podCreationTimestamp="2025-10-02 07:49:16 +0000 UTC" firstStartedPulling="2025-10-02 07:49:17.547422574 +0000 UTC m=+1948.887070979" lastFinishedPulling="2025-10-02 07:49:18.063631909 +0000 UTC m=+1949.403280324" observedRunningTime="2025-10-02 07:49:18.603629032 +0000 UTC m=+1949.943277457" watchObservedRunningTime="2025-10-02 07:49:18.613731128 +0000 UTC m=+1949.953379543" Oct 02 07:49:25 crc kubenswrapper[4829]: I1002 07:49:25.328702 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:49:25 crc kubenswrapper[4829]: I1002 07:49:25.329210 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:49:28 crc kubenswrapper[4829]: I1002 07:49:28.727531 4829 generic.go:334] "Generic (PLEG): container finished" podID="3ed1e719-596e-42fd-ac3f-b092996c9ca6" containerID="520e3e77500d848b6eea90ac2b75abf86a0afa3e939bf03770ed299b5faeea72" exitCode=0 Oct 02 07:49:28 crc kubenswrapper[4829]: I1002 07:49:28.727601 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" event={"ID":"3ed1e719-596e-42fd-ac3f-b092996c9ca6","Type":"ContainerDied","Data":"520e3e77500d848b6eea90ac2b75abf86a0afa3e939bf03770ed299b5faeea72"} Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.241446 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.284049 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-ssh-key\") pod \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.284330 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-inventory\") pod \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.284838 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx9hq\" (UniqueName: \"kubernetes.io/projected/3ed1e719-596e-42fd-ac3f-b092996c9ca6-kube-api-access-jx9hq\") pod \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\" (UID: \"3ed1e719-596e-42fd-ac3f-b092996c9ca6\") " Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.312446 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ed1e719-596e-42fd-ac3f-b092996c9ca6-kube-api-access-jx9hq" (OuterVolumeSpecName: "kube-api-access-jx9hq") pod "3ed1e719-596e-42fd-ac3f-b092996c9ca6" (UID: "3ed1e719-596e-42fd-ac3f-b092996c9ca6"). InnerVolumeSpecName "kube-api-access-jx9hq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.328264 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-inventory" (OuterVolumeSpecName: "inventory") pod "3ed1e719-596e-42fd-ac3f-b092996c9ca6" (UID: "3ed1e719-596e-42fd-ac3f-b092996c9ca6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.331197 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3ed1e719-596e-42fd-ac3f-b092996c9ca6" (UID: "3ed1e719-596e-42fd-ac3f-b092996c9ca6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.386694 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.386722 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3ed1e719-596e-42fd-ac3f-b092996c9ca6-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.386732 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx9hq\" (UniqueName: \"kubernetes.io/projected/3ed1e719-596e-42fd-ac3f-b092996c9ca6-kube-api-access-jx9hq\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.751439 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" event={"ID":"3ed1e719-596e-42fd-ac3f-b092996c9ca6","Type":"ContainerDied","Data":"12888828ec842ee18ef1a13d268ec18277d6272b8ca1e479812f8a1d7b84c8f1"} Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.751491 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12888828ec842ee18ef1a13d268ec18277d6272b8ca1e479812f8a1d7b84c8f1" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.751521 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2kgpj" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.844418 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87"] Oct 02 07:49:30 crc kubenswrapper[4829]: E1002 07:49:30.845033 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ed1e719-596e-42fd-ac3f-b092996c9ca6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.845058 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ed1e719-596e-42fd-ac3f-b092996c9ca6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.845402 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ed1e719-596e-42fd-ac3f-b092996c9ca6" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.846485 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.855128 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.855194 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.857552 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.861711 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.863024 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87"] Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.904362 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px5lz\" (UniqueName: \"kubernetes.io/projected/4d515775-5135-4f3f-aa21-fd1f648afb11-kube-api-access-px5lz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.904605 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:30 crc kubenswrapper[4829]: I1002 07:49:30.904647 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.006773 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px5lz\" (UniqueName: \"kubernetes.io/projected/4d515775-5135-4f3f-aa21-fd1f648afb11-kube-api-access-px5lz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.006985 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.007029 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.012408 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.015294 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.040822 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px5lz\" (UniqueName: \"kubernetes.io/projected/4d515775-5135-4f3f-aa21-fd1f648afb11-kube-api-access-px5lz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.185795 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.741410 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87"] Oct 02 07:49:31 crc kubenswrapper[4829]: I1002 07:49:31.763595 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" event={"ID":"4d515775-5135-4f3f-aa21-fd1f648afb11","Type":"ContainerStarted","Data":"460d1083b80fc84c15f4246e8ef92f9091b4f9d26ec2870369d8f24ade92479a"} Oct 02 07:49:32 crc kubenswrapper[4829]: I1002 07:49:32.779318 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" event={"ID":"4d515775-5135-4f3f-aa21-fd1f648afb11","Type":"ContainerStarted","Data":"f633f788aa9912d9b36d3d5681bebeb0e37f759b2724463ca9ee380bfb50c32e"} Oct 02 07:49:32 crc kubenswrapper[4829]: I1002 07:49:32.805604 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" podStartSLOduration=2.252859989 podStartE2EDuration="2.805566794s" podCreationTimestamp="2025-10-02 07:49:30 +0000 UTC" firstStartedPulling="2025-10-02 07:49:31.74881706 +0000 UTC m=+1963.088465475" lastFinishedPulling="2025-10-02 07:49:32.301523865 +0000 UTC m=+1963.641172280" observedRunningTime="2025-10-02 07:49:32.798203019 +0000 UTC m=+1964.137851474" watchObservedRunningTime="2025-10-02 07:49:32.805566794 +0000 UTC m=+1964.145215249" Oct 02 07:49:43 crc kubenswrapper[4829]: I1002 07:49:43.891212 4829 generic.go:334] "Generic (PLEG): container finished" podID="4d515775-5135-4f3f-aa21-fd1f648afb11" containerID="f633f788aa9912d9b36d3d5681bebeb0e37f759b2724463ca9ee380bfb50c32e" exitCode=0 Oct 02 07:49:43 crc kubenswrapper[4829]: I1002 07:49:43.891332 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" event={"ID":"4d515775-5135-4f3f-aa21-fd1f648afb11","Type":"ContainerDied","Data":"f633f788aa9912d9b36d3d5681bebeb0e37f759b2724463ca9ee380bfb50c32e"} Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.418577 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.537271 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-inventory\") pod \"4d515775-5135-4f3f-aa21-fd1f648afb11\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.537618 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px5lz\" (UniqueName: \"kubernetes.io/projected/4d515775-5135-4f3f-aa21-fd1f648afb11-kube-api-access-px5lz\") pod \"4d515775-5135-4f3f-aa21-fd1f648afb11\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.537870 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-ssh-key\") pod \"4d515775-5135-4f3f-aa21-fd1f648afb11\" (UID: \"4d515775-5135-4f3f-aa21-fd1f648afb11\") " Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.543292 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d515775-5135-4f3f-aa21-fd1f648afb11-kube-api-access-px5lz" (OuterVolumeSpecName: "kube-api-access-px5lz") pod "4d515775-5135-4f3f-aa21-fd1f648afb11" (UID: "4d515775-5135-4f3f-aa21-fd1f648afb11"). InnerVolumeSpecName "kube-api-access-px5lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.569213 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-inventory" (OuterVolumeSpecName: "inventory") pod "4d515775-5135-4f3f-aa21-fd1f648afb11" (UID: "4d515775-5135-4f3f-aa21-fd1f648afb11"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.577848 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d515775-5135-4f3f-aa21-fd1f648afb11" (UID: "4d515775-5135-4f3f-aa21-fd1f648afb11"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.641312 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.641345 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px5lz\" (UniqueName: \"kubernetes.io/projected/4d515775-5135-4f3f-aa21-fd1f648afb11-kube-api-access-px5lz\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.641356 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d515775-5135-4f3f-aa21-fd1f648afb11-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.930650 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" event={"ID":"4d515775-5135-4f3f-aa21-fd1f648afb11","Type":"ContainerDied","Data":"460d1083b80fc84c15f4246e8ef92f9091b4f9d26ec2870369d8f24ade92479a"} Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.930728 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="460d1083b80fc84c15f4246e8ef92f9091b4f9d26ec2870369d8f24ade92479a" Oct 02 07:49:45 crc kubenswrapper[4829]: I1002 07:49:45.930803 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.078951 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz"] Oct 02 07:49:46 crc kubenswrapper[4829]: E1002 07:49:46.079915 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d515775-5135-4f3f-aa21-fd1f648afb11" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.079961 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d515775-5135-4f3f-aa21-fd1f648afb11" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.080540 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d515775-5135-4f3f-aa21-fd1f648afb11" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.082113 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.087086 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.087270 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.087454 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.087672 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.088114 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.088131 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.089039 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.089286 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.098626 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz"] Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155052 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155129 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155282 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155322 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155411 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155464 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155486 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155508 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155574 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155610 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155633 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155657 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4pbr\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-kube-api-access-c4pbr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155686 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.155703 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257238 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257296 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257318 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257362 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257402 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257423 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.257441 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258059 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258105 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258129 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258155 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4pbr\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-kube-api-access-c4pbr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258185 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258206 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.258276 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.265043 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.265592 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.266066 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.266085 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.266217 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.266321 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.268060 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.268122 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.268553 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.269319 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.269562 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.269896 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.271691 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.276998 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4pbr\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-kube-api-access-c4pbr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-7krkz\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.408421 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:49:46 crc kubenswrapper[4829]: I1002 07:49:46.929396 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz"] Oct 02 07:49:47 crc kubenswrapper[4829]: I1002 07:49:47.950630 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" event={"ID":"28912442-1f56-461e-9d26-18227d1f2386","Type":"ContainerStarted","Data":"2e7e6f2d89082de19964b29598508254230c8bd43e88bdd582e46fc8b882fb61"} Oct 02 07:49:47 crc kubenswrapper[4829]: I1002 07:49:47.950682 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" event={"ID":"28912442-1f56-461e-9d26-18227d1f2386","Type":"ContainerStarted","Data":"e615a5e242384800c9b89afaee3887e1fff33658acd3982415b1bc06f9537ce8"} Oct 02 07:49:47 crc kubenswrapper[4829]: I1002 07:49:47.971451 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" podStartSLOduration=1.284198118 podStartE2EDuration="1.971422568s" podCreationTimestamp="2025-10-02 07:49:46 +0000 UTC" firstStartedPulling="2025-10-02 07:49:46.943604882 +0000 UTC m=+1978.283253287" lastFinishedPulling="2025-10-02 07:49:47.630829332 +0000 UTC m=+1978.970477737" observedRunningTime="2025-10-02 07:49:47.966119543 +0000 UTC m=+1979.305767948" watchObservedRunningTime="2025-10-02 07:49:47.971422568 +0000 UTC m=+1979.311070993" Oct 02 07:49:55 crc kubenswrapper[4829]: I1002 07:49:55.329166 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:49:55 crc kubenswrapper[4829]: I1002 07:49:55.329796 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:50:25 crc kubenswrapper[4829]: I1002 07:50:25.329508 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:50:25 crc kubenswrapper[4829]: I1002 07:50:25.330136 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:50:25 crc kubenswrapper[4829]: I1002 07:50:25.330202 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:50:25 crc kubenswrapper[4829]: I1002 07:50:25.331404 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"46cd74959e46dfb078660ee3089447f9db6cde42ac4b590e00c9b5fae637513e"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:50:25 crc kubenswrapper[4829]: I1002 07:50:25.331508 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://46cd74959e46dfb078660ee3089447f9db6cde42ac4b590e00c9b5fae637513e" gracePeriod=600 Oct 02 07:50:26 crc kubenswrapper[4829]: I1002 07:50:26.400571 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="46cd74959e46dfb078660ee3089447f9db6cde42ac4b590e00c9b5fae637513e" exitCode=0 Oct 02 07:50:26 crc kubenswrapper[4829]: I1002 07:50:26.400647 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"46cd74959e46dfb078660ee3089447f9db6cde42ac4b590e00c9b5fae637513e"} Oct 02 07:50:26 crc kubenswrapper[4829]: I1002 07:50:26.401147 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4"} Oct 02 07:50:26 crc kubenswrapper[4829]: I1002 07:50:26.401169 4829 scope.go:117] "RemoveContainer" containerID="1f0a239e07e27723eb91f4ef10d21e7b232bde748e2397687a328dc2979c7e31" Oct 02 07:50:30 crc kubenswrapper[4829]: I1002 07:50:30.458930 4829 generic.go:334] "Generic (PLEG): container finished" podID="28912442-1f56-461e-9d26-18227d1f2386" containerID="2e7e6f2d89082de19964b29598508254230c8bd43e88bdd582e46fc8b882fb61" exitCode=0 Oct 02 07:50:30 crc kubenswrapper[4829]: I1002 07:50:30.459290 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" event={"ID":"28912442-1f56-461e-9d26-18227d1f2386","Type":"ContainerDied","Data":"2e7e6f2d89082de19964b29598508254230c8bd43e88bdd582e46fc8b882fb61"} Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.921497 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932179 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932260 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ovn-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932298 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-bootstrap-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932319 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-libvirt-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932336 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-nova-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932396 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-inventory\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932417 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-repo-setup-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932504 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-telemetry-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932523 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932546 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932597 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ssh-key\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932643 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-ovn-default-certs-0\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932705 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4pbr\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-kube-api-access-c4pbr\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.932765 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-neutron-metadata-combined-ca-bundle\") pod \"28912442-1f56-461e-9d26-18227d1f2386\" (UID: \"28912442-1f56-461e-9d26-18227d1f2386\") " Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.940153 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.940702 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.941116 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.941830 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.941979 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.943079 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.944028 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-kube-api-access-c4pbr" (OuterVolumeSpecName: "kube-api-access-c4pbr") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "kube-api-access-c4pbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.947012 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.948674 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.952909 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.962718 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.971139 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.983431 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:31 crc kubenswrapper[4829]: I1002 07:50:31.984518 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-inventory" (OuterVolumeSpecName: "inventory") pod "28912442-1f56-461e-9d26-18227d1f2386" (UID: "28912442-1f56-461e-9d26-18227d1f2386"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036781 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4pbr\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-kube-api-access-c4pbr\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036836 4829 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036860 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036882 4829 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036901 4829 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036917 4829 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036933 4829 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036950 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036965 4829 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036980 4829 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.036997 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.037014 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.037032 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28912442-1f56-461e-9d26-18227d1f2386-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.037048 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/28912442-1f56-461e-9d26-18227d1f2386-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.487746 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" event={"ID":"28912442-1f56-461e-9d26-18227d1f2386","Type":"ContainerDied","Data":"e615a5e242384800c9b89afaee3887e1fff33658acd3982415b1bc06f9537ce8"} Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.488412 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e615a5e242384800c9b89afaee3887e1fff33658acd3982415b1bc06f9537ce8" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.487840 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-7krkz" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.634163 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v"] Oct 02 07:50:32 crc kubenswrapper[4829]: E1002 07:50:32.634598 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28912442-1f56-461e-9d26-18227d1f2386" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.634615 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="28912442-1f56-461e-9d26-18227d1f2386" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.634855 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="28912442-1f56-461e-9d26-18227d1f2386" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.635527 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.647863 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.647952 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.648251 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.648489 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.648562 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.650371 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.650487 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2qhz\" (UniqueName: \"kubernetes.io/projected/adb0b737-9122-4893-8b4d-2111d18c2a4d-kube-api-access-p2qhz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.650574 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.650652 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.650706 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.655447 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v"] Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.752664 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.752741 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.752811 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.752926 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2qhz\" (UniqueName: \"kubernetes.io/projected/adb0b737-9122-4893-8b4d-2111d18c2a4d-kube-api-access-p2qhz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.753007 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.753943 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.756253 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.762077 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.762309 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.776019 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2qhz\" (UniqueName: \"kubernetes.io/projected/adb0b737-9122-4893-8b4d-2111d18c2a4d-kube-api-access-p2qhz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xsv2v\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:32 crc kubenswrapper[4829]: I1002 07:50:32.951792 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:50:33 crc kubenswrapper[4829]: I1002 07:50:33.527071 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v"] Oct 02 07:50:33 crc kubenswrapper[4829]: W1002 07:50:33.532064 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podadb0b737_9122_4893_8b4d_2111d18c2a4d.slice/crio-5d32447c22d24651386ab5272c5c5aabd9b7260d98ed86ccbda36aa6af5eb93e WatchSource:0}: Error finding container 5d32447c22d24651386ab5272c5c5aabd9b7260d98ed86ccbda36aa6af5eb93e: Status 404 returned error can't find the container with id 5d32447c22d24651386ab5272c5c5aabd9b7260d98ed86ccbda36aa6af5eb93e Oct 02 07:50:34 crc kubenswrapper[4829]: I1002 07:50:34.510914 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" event={"ID":"adb0b737-9122-4893-8b4d-2111d18c2a4d","Type":"ContainerStarted","Data":"73a022491b4e7d78b62afe5b7a549c61e365343ddadf1e5dea03ac3de05cbc74"} Oct 02 07:50:34 crc kubenswrapper[4829]: I1002 07:50:34.511806 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" event={"ID":"adb0b737-9122-4893-8b4d-2111d18c2a4d","Type":"ContainerStarted","Data":"5d32447c22d24651386ab5272c5c5aabd9b7260d98ed86ccbda36aa6af5eb93e"} Oct 02 07:50:34 crc kubenswrapper[4829]: I1002 07:50:34.532520 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" podStartSLOduration=2.111545609 podStartE2EDuration="2.532497012s" podCreationTimestamp="2025-10-02 07:50:32 +0000 UTC" firstStartedPulling="2025-10-02 07:50:33.533852301 +0000 UTC m=+2024.873500706" lastFinishedPulling="2025-10-02 07:50:33.954803704 +0000 UTC m=+2025.294452109" observedRunningTime="2025-10-02 07:50:34.5276436 +0000 UTC m=+2025.867292015" watchObservedRunningTime="2025-10-02 07:50:34.532497012 +0000 UTC m=+2025.872145427" Oct 02 07:51:46 crc kubenswrapper[4829]: I1002 07:51:46.315808 4829 generic.go:334] "Generic (PLEG): container finished" podID="adb0b737-9122-4893-8b4d-2111d18c2a4d" containerID="73a022491b4e7d78b62afe5b7a549c61e365343ddadf1e5dea03ac3de05cbc74" exitCode=0 Oct 02 07:51:46 crc kubenswrapper[4829]: I1002 07:51:46.315897 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" event={"ID":"adb0b737-9122-4893-8b4d-2111d18c2a4d","Type":"ContainerDied","Data":"73a022491b4e7d78b62afe5b7a549c61e365343ddadf1e5dea03ac3de05cbc74"} Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.826656 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.958785 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2qhz\" (UniqueName: \"kubernetes.io/projected/adb0b737-9122-4893-8b4d-2111d18c2a4d-kube-api-access-p2qhz\") pod \"adb0b737-9122-4893-8b4d-2111d18c2a4d\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.959082 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ssh-key\") pod \"adb0b737-9122-4893-8b4d-2111d18c2a4d\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.959146 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-inventory\") pod \"adb0b737-9122-4893-8b4d-2111d18c2a4d\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.959356 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovncontroller-config-0\") pod \"adb0b737-9122-4893-8b4d-2111d18c2a4d\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.959456 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovn-combined-ca-bundle\") pod \"adb0b737-9122-4893-8b4d-2111d18c2a4d\" (UID: \"adb0b737-9122-4893-8b4d-2111d18c2a4d\") " Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.973213 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adb0b737-9122-4893-8b4d-2111d18c2a4d-kube-api-access-p2qhz" (OuterVolumeSpecName: "kube-api-access-p2qhz") pod "adb0b737-9122-4893-8b4d-2111d18c2a4d" (UID: "adb0b737-9122-4893-8b4d-2111d18c2a4d"). InnerVolumeSpecName "kube-api-access-p2qhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:51:47 crc kubenswrapper[4829]: I1002 07:51:47.974579 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "adb0b737-9122-4893-8b4d-2111d18c2a4d" (UID: "adb0b737-9122-4893-8b4d-2111d18c2a4d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.004308 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "adb0b737-9122-4893-8b4d-2111d18c2a4d" (UID: "adb0b737-9122-4893-8b4d-2111d18c2a4d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.005079 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-inventory" (OuterVolumeSpecName: "inventory") pod "adb0b737-9122-4893-8b4d-2111d18c2a4d" (UID: "adb0b737-9122-4893-8b4d-2111d18c2a4d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.016350 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "adb0b737-9122-4893-8b4d-2111d18c2a4d" (UID: "adb0b737-9122-4893-8b4d-2111d18c2a4d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.061686 4829 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.061735 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2qhz\" (UniqueName: \"kubernetes.io/projected/adb0b737-9122-4893-8b4d-2111d18c2a4d-kube-api-access-p2qhz\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.061754 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.061772 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adb0b737-9122-4893-8b4d-2111d18c2a4d-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.061791 4829 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adb0b737-9122-4893-8b4d-2111d18c2a4d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.349546 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" event={"ID":"adb0b737-9122-4893-8b4d-2111d18c2a4d","Type":"ContainerDied","Data":"5d32447c22d24651386ab5272c5c5aabd9b7260d98ed86ccbda36aa6af5eb93e"} Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.349829 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d32447c22d24651386ab5272c5c5aabd9b7260d98ed86ccbda36aa6af5eb93e" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.349943 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xsv2v" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.547455 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc"] Oct 02 07:51:48 crc kubenswrapper[4829]: E1002 07:51:48.548192 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adb0b737-9122-4893-8b4d-2111d18c2a4d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.548209 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="adb0b737-9122-4893-8b4d-2111d18c2a4d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.548409 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="adb0b737-9122-4893-8b4d-2111d18c2a4d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.549062 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.556197 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.556559 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.556607 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.556947 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.556996 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.561818 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.580772 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.580871 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.580991 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.581052 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.581129 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.581185 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnjx7\" (UniqueName: \"kubernetes.io/projected/d0041618-1a4e-474f-8570-ccf383c6cff9-kube-api-access-pnjx7\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.582774 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc"] Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.682151 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.682449 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.682607 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.682700 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnjx7\" (UniqueName: \"kubernetes.io/projected/d0041618-1a4e-474f-8570-ccf383c6cff9-kube-api-access-pnjx7\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.682787 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.682882 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.686423 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.687150 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.687262 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.688157 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.689102 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.706995 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnjx7\" (UniqueName: \"kubernetes.io/projected/d0041618-1a4e-474f-8570-ccf383c6cff9-kube-api-access-pnjx7\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:48 crc kubenswrapper[4829]: I1002 07:51:48.887802 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:51:49 crc kubenswrapper[4829]: I1002 07:51:49.536419 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc"] Oct 02 07:51:50 crc kubenswrapper[4829]: I1002 07:51:50.039411 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:51:50 crc kubenswrapper[4829]: I1002 07:51:50.375163 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" event={"ID":"d0041618-1a4e-474f-8570-ccf383c6cff9","Type":"ContainerStarted","Data":"08efaf602d57927fee2ae2c9e3cebc4dc2c5cc53fbdb81615e378b22931241ff"} Oct 02 07:51:50 crc kubenswrapper[4829]: I1002 07:51:50.375277 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" event={"ID":"d0041618-1a4e-474f-8570-ccf383c6cff9","Type":"ContainerStarted","Data":"2b853a25e7b83538f951dd48a7474359a8d0da4bf17330608167169fb7f42182"} Oct 02 07:51:50 crc kubenswrapper[4829]: I1002 07:51:50.402867 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" podStartSLOduration=1.903689991 podStartE2EDuration="2.402833406s" podCreationTimestamp="2025-10-02 07:51:48 +0000 UTC" firstStartedPulling="2025-10-02 07:51:49.536559046 +0000 UTC m=+2100.876207451" lastFinishedPulling="2025-10-02 07:51:50.035702431 +0000 UTC m=+2101.375350866" observedRunningTime="2025-10-02 07:51:50.397981814 +0000 UTC m=+2101.737630269" watchObservedRunningTime="2025-10-02 07:51:50.402833406 +0000 UTC m=+2101.742481861" Oct 02 07:52:25 crc kubenswrapper[4829]: I1002 07:52:25.329567 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:52:25 crc kubenswrapper[4829]: I1002 07:52:25.330258 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:52:45 crc kubenswrapper[4829]: I1002 07:52:45.010305 4829 generic.go:334] "Generic (PLEG): container finished" podID="d0041618-1a4e-474f-8570-ccf383c6cff9" containerID="08efaf602d57927fee2ae2c9e3cebc4dc2c5cc53fbdb81615e378b22931241ff" exitCode=0 Oct 02 07:52:45 crc kubenswrapper[4829]: I1002 07:52:45.010414 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" event={"ID":"d0041618-1a4e-474f-8570-ccf383c6cff9","Type":"ContainerDied","Data":"08efaf602d57927fee2ae2c9e3cebc4dc2c5cc53fbdb81615e378b22931241ff"} Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.575863 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.647975 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-nova-metadata-neutron-config-0\") pod \"d0041618-1a4e-474f-8570-ccf383c6cff9\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.648023 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-ovn-metadata-agent-neutron-config-0\") pod \"d0041618-1a4e-474f-8570-ccf383c6cff9\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.648062 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-metadata-combined-ca-bundle\") pod \"d0041618-1a4e-474f-8570-ccf383c6cff9\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.648089 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnjx7\" (UniqueName: \"kubernetes.io/projected/d0041618-1a4e-474f-8570-ccf383c6cff9-kube-api-access-pnjx7\") pod \"d0041618-1a4e-474f-8570-ccf383c6cff9\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.648202 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-inventory\") pod \"d0041618-1a4e-474f-8570-ccf383c6cff9\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.648266 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-ssh-key\") pod \"d0041618-1a4e-474f-8570-ccf383c6cff9\" (UID: \"d0041618-1a4e-474f-8570-ccf383c6cff9\") " Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.653313 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0041618-1a4e-474f-8570-ccf383c6cff9-kube-api-access-pnjx7" (OuterVolumeSpecName: "kube-api-access-pnjx7") pod "d0041618-1a4e-474f-8570-ccf383c6cff9" (UID: "d0041618-1a4e-474f-8570-ccf383c6cff9"). InnerVolumeSpecName "kube-api-access-pnjx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.654641 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d0041618-1a4e-474f-8570-ccf383c6cff9" (UID: "d0041618-1a4e-474f-8570-ccf383c6cff9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.674967 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-inventory" (OuterVolumeSpecName: "inventory") pod "d0041618-1a4e-474f-8570-ccf383c6cff9" (UID: "d0041618-1a4e-474f-8570-ccf383c6cff9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.679926 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d0041618-1a4e-474f-8570-ccf383c6cff9" (UID: "d0041618-1a4e-474f-8570-ccf383c6cff9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.681146 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "d0041618-1a4e-474f-8570-ccf383c6cff9" (UID: "d0041618-1a4e-474f-8570-ccf383c6cff9"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.701645 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "d0041618-1a4e-474f-8570-ccf383c6cff9" (UID: "d0041618-1a4e-474f-8570-ccf383c6cff9"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.750602 4829 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.750648 4829 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.750669 4829 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.750688 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnjx7\" (UniqueName: \"kubernetes.io/projected/d0041618-1a4e-474f-8570-ccf383c6cff9-kube-api-access-pnjx7\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.750704 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:46 crc kubenswrapper[4829]: I1002 07:52:46.750718 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0041618-1a4e-474f-8570-ccf383c6cff9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.038411 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" event={"ID":"d0041618-1a4e-474f-8570-ccf383c6cff9","Type":"ContainerDied","Data":"2b853a25e7b83538f951dd48a7474359a8d0da4bf17330608167169fb7f42182"} Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.038470 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.038480 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b853a25e7b83538f951dd48a7474359a8d0da4bf17330608167169fb7f42182" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.161539 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd"] Oct 02 07:52:47 crc kubenswrapper[4829]: E1002 07:52:47.162042 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0041618-1a4e-474f-8570-ccf383c6cff9" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.162067 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0041618-1a4e-474f-8570-ccf383c6cff9" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.162369 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0041618-1a4e-474f-8570-ccf383c6cff9" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.163100 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.169096 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.169131 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.169516 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.169600 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.175799 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.185056 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd"] Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.260346 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.260520 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbwvl\" (UniqueName: \"kubernetes.io/projected/9dfa4494-c2fb-415c-a38a-79afbf13be81-kube-api-access-qbwvl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.260733 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.261249 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.261393 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.363188 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.363276 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbwvl\" (UniqueName: \"kubernetes.io/projected/9dfa4494-c2fb-415c-a38a-79afbf13be81-kube-api-access-qbwvl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.363316 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.364061 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.364390 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.369131 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.369186 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.369830 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.370610 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.390586 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbwvl\" (UniqueName: \"kubernetes.io/projected/9dfa4494-c2fb-415c-a38a-79afbf13be81-kube-api-access-qbwvl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-94ftd\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:47 crc kubenswrapper[4829]: I1002 07:52:47.486553 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:52:48 crc kubenswrapper[4829]: I1002 07:52:48.099876 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd"] Oct 02 07:52:48 crc kubenswrapper[4829]: W1002 07:52:48.103883 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9dfa4494_c2fb_415c_a38a_79afbf13be81.slice/crio-3b7c1346cfa4b6414dc016631fc30317cb921b05afcf87ca295b50eaee52d681 WatchSource:0}: Error finding container 3b7c1346cfa4b6414dc016631fc30317cb921b05afcf87ca295b50eaee52d681: Status 404 returned error can't find the container with id 3b7c1346cfa4b6414dc016631fc30317cb921b05afcf87ca295b50eaee52d681 Oct 02 07:52:49 crc kubenswrapper[4829]: I1002 07:52:49.062139 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" event={"ID":"9dfa4494-c2fb-415c-a38a-79afbf13be81","Type":"ContainerStarted","Data":"822162a5f255552e31eecb4e42999bc2b21e542127a0420bb2d823543a771220"} Oct 02 07:52:49 crc kubenswrapper[4829]: I1002 07:52:49.062818 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" event={"ID":"9dfa4494-c2fb-415c-a38a-79afbf13be81","Type":"ContainerStarted","Data":"3b7c1346cfa4b6414dc016631fc30317cb921b05afcf87ca295b50eaee52d681"} Oct 02 07:52:49 crc kubenswrapper[4829]: I1002 07:52:49.085339 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" podStartSLOduration=1.408056064 podStartE2EDuration="2.085310398s" podCreationTimestamp="2025-10-02 07:52:47 +0000 UTC" firstStartedPulling="2025-10-02 07:52:48.106605517 +0000 UTC m=+2159.446253942" lastFinishedPulling="2025-10-02 07:52:48.783859831 +0000 UTC m=+2160.123508276" observedRunningTime="2025-10-02 07:52:49.080892888 +0000 UTC m=+2160.420541323" watchObservedRunningTime="2025-10-02 07:52:49.085310398 +0000 UTC m=+2160.424958853" Oct 02 07:52:55 crc kubenswrapper[4829]: I1002 07:52:55.329649 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:52:55 crc kubenswrapper[4829]: I1002 07:52:55.330348 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.329123 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.329640 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.329690 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.330486 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.330555 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" gracePeriod=600 Oct 02 07:53:25 crc kubenswrapper[4829]: E1002 07:53:25.373718 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60aa0a81_354e_46f1_ab0c_b1eb386974a6.slice/crio-conmon-1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4.scope\": RecentStats: unable to find data in memory cache]" Oct 02 07:53:25 crc kubenswrapper[4829]: E1002 07:53:25.457422 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.479529 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" exitCode=0 Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.479881 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4"} Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.479916 4829 scope.go:117] "RemoveContainer" containerID="46cd74959e46dfb078660ee3089447f9db6cde42ac4b590e00c9b5fae637513e" Oct 02 07:53:25 crc kubenswrapper[4829]: I1002 07:53:25.480651 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:53:25 crc kubenswrapper[4829]: E1002 07:53:25.480915 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:53:36 crc kubenswrapper[4829]: I1002 07:53:36.463526 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:53:36 crc kubenswrapper[4829]: E1002 07:53:36.464641 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:53:47 crc kubenswrapper[4829]: I1002 07:53:47.461777 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:53:47 crc kubenswrapper[4829]: E1002 07:53:47.462589 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:54:02 crc kubenswrapper[4829]: I1002 07:54:02.460398 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:54:02 crc kubenswrapper[4829]: E1002 07:54:02.461079 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:54:17 crc kubenswrapper[4829]: I1002 07:54:17.461462 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:54:17 crc kubenswrapper[4829]: E1002 07:54:17.462293 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:54:28 crc kubenswrapper[4829]: I1002 07:54:28.461479 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:54:28 crc kubenswrapper[4829]: E1002 07:54:28.462305 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:54:41 crc kubenswrapper[4829]: I1002 07:54:41.461576 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:54:41 crc kubenswrapper[4829]: E1002 07:54:41.462594 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:54:56 crc kubenswrapper[4829]: I1002 07:54:56.461455 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:54:56 crc kubenswrapper[4829]: E1002 07:54:56.462519 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:55:10 crc kubenswrapper[4829]: I1002 07:55:10.460994 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:55:10 crc kubenswrapper[4829]: E1002 07:55:10.461946 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.346283 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qfh5b"] Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.352470 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.365953 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qfh5b"] Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.413298 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-catalog-content\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.413547 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2dlx\" (UniqueName: \"kubernetes.io/projected/05538b82-50bd-47c4-91a0-366cd3fd4f24-kube-api-access-m2dlx\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.413748 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-utilities\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.516574 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2dlx\" (UniqueName: \"kubernetes.io/projected/05538b82-50bd-47c4-91a0-366cd3fd4f24-kube-api-access-m2dlx\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.516726 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-utilities\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.516791 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-catalog-content\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.517334 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-catalog-content\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.517341 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-utilities\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.537846 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2dlx\" (UniqueName: \"kubernetes.io/projected/05538b82-50bd-47c4-91a0-366cd3fd4f24-kube-api-access-m2dlx\") pod \"community-operators-qfh5b\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:22 crc kubenswrapper[4829]: I1002 07:55:22.732431 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:23 crc kubenswrapper[4829]: I1002 07:55:23.239505 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qfh5b"] Oct 02 07:55:23 crc kubenswrapper[4829]: W1002 07:55:23.245819 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05538b82_50bd_47c4_91a0_366cd3fd4f24.slice/crio-6f9a2456853b0cfdcb06137a984dc468bfbf997a3aaa2ee02779faffbb087ad1 WatchSource:0}: Error finding container 6f9a2456853b0cfdcb06137a984dc468bfbf997a3aaa2ee02779faffbb087ad1: Status 404 returned error can't find the container with id 6f9a2456853b0cfdcb06137a984dc468bfbf997a3aaa2ee02779faffbb087ad1 Oct 02 07:55:23 crc kubenswrapper[4829]: I1002 07:55:23.785276 4829 generic.go:334] "Generic (PLEG): container finished" podID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerID="14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a" exitCode=0 Oct 02 07:55:23 crc kubenswrapper[4829]: I1002 07:55:23.785366 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerDied","Data":"14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a"} Oct 02 07:55:23 crc kubenswrapper[4829]: I1002 07:55:23.785561 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerStarted","Data":"6f9a2456853b0cfdcb06137a984dc468bfbf997a3aaa2ee02779faffbb087ad1"} Oct 02 07:55:23 crc kubenswrapper[4829]: I1002 07:55:23.789147 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 07:55:24 crc kubenswrapper[4829]: I1002 07:55:24.461265 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:55:24 crc kubenswrapper[4829]: E1002 07:55:24.461477 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:55:24 crc kubenswrapper[4829]: I1002 07:55:24.796173 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerStarted","Data":"2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889"} Oct 02 07:55:26 crc kubenswrapper[4829]: I1002 07:55:26.815045 4829 generic.go:334] "Generic (PLEG): container finished" podID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerID="2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889" exitCode=0 Oct 02 07:55:26 crc kubenswrapper[4829]: I1002 07:55:26.815416 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerDied","Data":"2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889"} Oct 02 07:55:27 crc kubenswrapper[4829]: I1002 07:55:27.830001 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerStarted","Data":"db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232"} Oct 02 07:55:32 crc kubenswrapper[4829]: I1002 07:55:32.732684 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:32 crc kubenswrapper[4829]: I1002 07:55:32.733673 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:32 crc kubenswrapper[4829]: I1002 07:55:32.802687 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:32 crc kubenswrapper[4829]: I1002 07:55:32.842374 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qfh5b" podStartSLOduration=7.410099601 podStartE2EDuration="10.842347337s" podCreationTimestamp="2025-10-02 07:55:22 +0000 UTC" firstStartedPulling="2025-10-02 07:55:23.788844437 +0000 UTC m=+2315.128492842" lastFinishedPulling="2025-10-02 07:55:27.221092143 +0000 UTC m=+2318.560740578" observedRunningTime="2025-10-02 07:55:27.860656541 +0000 UTC m=+2319.200304956" watchObservedRunningTime="2025-10-02 07:55:32.842347337 +0000 UTC m=+2324.181995772" Oct 02 07:55:32 crc kubenswrapper[4829]: I1002 07:55:32.954276 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:33 crc kubenswrapper[4829]: I1002 07:55:33.049073 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qfh5b"] Oct 02 07:55:34 crc kubenswrapper[4829]: I1002 07:55:34.914493 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qfh5b" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="registry-server" containerID="cri-o://db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232" gracePeriod=2 Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.382141 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.582009 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-catalog-content\") pod \"05538b82-50bd-47c4-91a0-366cd3fd4f24\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.582251 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-utilities\") pod \"05538b82-50bd-47c4-91a0-366cd3fd4f24\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.582376 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2dlx\" (UniqueName: \"kubernetes.io/projected/05538b82-50bd-47c4-91a0-366cd3fd4f24-kube-api-access-m2dlx\") pod \"05538b82-50bd-47c4-91a0-366cd3fd4f24\" (UID: \"05538b82-50bd-47c4-91a0-366cd3fd4f24\") " Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.583104 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-utilities" (OuterVolumeSpecName: "utilities") pod "05538b82-50bd-47c4-91a0-366cd3fd4f24" (UID: "05538b82-50bd-47c4-91a0-366cd3fd4f24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.590211 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05538b82-50bd-47c4-91a0-366cd3fd4f24-kube-api-access-m2dlx" (OuterVolumeSpecName: "kube-api-access-m2dlx") pod "05538b82-50bd-47c4-91a0-366cd3fd4f24" (UID: "05538b82-50bd-47c4-91a0-366cd3fd4f24"). InnerVolumeSpecName "kube-api-access-m2dlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.639903 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05538b82-50bd-47c4-91a0-366cd3fd4f24" (UID: "05538b82-50bd-47c4-91a0-366cd3fd4f24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.685396 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2dlx\" (UniqueName: \"kubernetes.io/projected/05538b82-50bd-47c4-91a0-366cd3fd4f24-kube-api-access-m2dlx\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.685437 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.685448 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05538b82-50bd-47c4-91a0-366cd3fd4f24-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.926435 4829 generic.go:334] "Generic (PLEG): container finished" podID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerID="db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232" exitCode=0 Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.926488 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerDied","Data":"db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232"} Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.926495 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfh5b" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.926523 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfh5b" event={"ID":"05538b82-50bd-47c4-91a0-366cd3fd4f24","Type":"ContainerDied","Data":"6f9a2456853b0cfdcb06137a984dc468bfbf997a3aaa2ee02779faffbb087ad1"} Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.926544 4829 scope.go:117] "RemoveContainer" containerID="db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.946542 4829 scope.go:117] "RemoveContainer" containerID="2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889" Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.967852 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qfh5b"] Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.975099 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qfh5b"] Oct 02 07:55:35 crc kubenswrapper[4829]: I1002 07:55:35.991077 4829 scope.go:117] "RemoveContainer" containerID="14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.026098 4829 scope.go:117] "RemoveContainer" containerID="db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232" Oct 02 07:55:36 crc kubenswrapper[4829]: E1002 07:55:36.026568 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232\": container with ID starting with db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232 not found: ID does not exist" containerID="db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.026610 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232"} err="failed to get container status \"db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232\": rpc error: code = NotFound desc = could not find container \"db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232\": container with ID starting with db291ce7abc729283b92bb72f9de7bc7f034841d077109662b7470cbc83e9232 not found: ID does not exist" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.026635 4829 scope.go:117] "RemoveContainer" containerID="2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889" Oct 02 07:55:36 crc kubenswrapper[4829]: E1002 07:55:36.026962 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889\": container with ID starting with 2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889 not found: ID does not exist" containerID="2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.027003 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889"} err="failed to get container status \"2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889\": rpc error: code = NotFound desc = could not find container \"2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889\": container with ID starting with 2a865811311591172cabeb0b1bfaeb54b3f2708ea18449ca98207795af1ff889 not found: ID does not exist" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.027030 4829 scope.go:117] "RemoveContainer" containerID="14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a" Oct 02 07:55:36 crc kubenswrapper[4829]: E1002 07:55:36.027330 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a\": container with ID starting with 14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a not found: ID does not exist" containerID="14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.027353 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a"} err="failed to get container status \"14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a\": rpc error: code = NotFound desc = could not find container \"14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a\": container with ID starting with 14fecfd43ba650e00a4e0d3f55e86324d7ad0cbadfdba87e4b0b73e7a5d24a4a not found: ID does not exist" Oct 02 07:55:36 crc kubenswrapper[4829]: I1002 07:55:36.461906 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:55:36 crc kubenswrapper[4829]: E1002 07:55:36.462611 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:55:37 crc kubenswrapper[4829]: I1002 07:55:37.480741 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" path="/var/lib/kubelet/pods/05538b82-50bd-47c4-91a0-366cd3fd4f24/volumes" Oct 02 07:55:49 crc kubenswrapper[4829]: I1002 07:55:49.466730 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:55:49 crc kubenswrapper[4829]: E1002 07:55:49.468022 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:56:02 crc kubenswrapper[4829]: I1002 07:56:02.460594 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:56:02 crc kubenswrapper[4829]: E1002 07:56:02.461393 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:56:15 crc kubenswrapper[4829]: I1002 07:56:15.461328 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:56:15 crc kubenswrapper[4829]: E1002 07:56:15.462595 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:56:30 crc kubenswrapper[4829]: I1002 07:56:30.461128 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:56:30 crc kubenswrapper[4829]: E1002 07:56:30.463961 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:56:45 crc kubenswrapper[4829]: I1002 07:56:45.461378 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:56:45 crc kubenswrapper[4829]: E1002 07:56:45.462183 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:57:00 crc kubenswrapper[4829]: I1002 07:57:00.461057 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:57:00 crc kubenswrapper[4829]: E1002 07:57:00.461917 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:57:13 crc kubenswrapper[4829]: I1002 07:57:13.461773 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:57:13 crc kubenswrapper[4829]: E1002 07:57:13.463057 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:57:23 crc kubenswrapper[4829]: I1002 07:57:23.069085 4829 generic.go:334] "Generic (PLEG): container finished" podID="9dfa4494-c2fb-415c-a38a-79afbf13be81" containerID="822162a5f255552e31eecb4e42999bc2b21e542127a0420bb2d823543a771220" exitCode=0 Oct 02 07:57:23 crc kubenswrapper[4829]: I1002 07:57:23.069647 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" event={"ID":"9dfa4494-c2fb-415c-a38a-79afbf13be81","Type":"ContainerDied","Data":"822162a5f255552e31eecb4e42999bc2b21e542127a0420bb2d823543a771220"} Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.517342 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.681104 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-ssh-key\") pod \"9dfa4494-c2fb-415c-a38a-79afbf13be81\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.681193 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-combined-ca-bundle\") pod \"9dfa4494-c2fb-415c-a38a-79afbf13be81\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.681958 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbwvl\" (UniqueName: \"kubernetes.io/projected/9dfa4494-c2fb-415c-a38a-79afbf13be81-kube-api-access-qbwvl\") pod \"9dfa4494-c2fb-415c-a38a-79afbf13be81\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.681991 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-secret-0\") pod \"9dfa4494-c2fb-415c-a38a-79afbf13be81\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.682158 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-inventory\") pod \"9dfa4494-c2fb-415c-a38a-79afbf13be81\" (UID: \"9dfa4494-c2fb-415c-a38a-79afbf13be81\") " Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.687060 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dfa4494-c2fb-415c-a38a-79afbf13be81-kube-api-access-qbwvl" (OuterVolumeSpecName: "kube-api-access-qbwvl") pod "9dfa4494-c2fb-415c-a38a-79afbf13be81" (UID: "9dfa4494-c2fb-415c-a38a-79afbf13be81"). InnerVolumeSpecName "kube-api-access-qbwvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.690304 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9dfa4494-c2fb-415c-a38a-79afbf13be81" (UID: "9dfa4494-c2fb-415c-a38a-79afbf13be81"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.717842 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-inventory" (OuterVolumeSpecName: "inventory") pod "9dfa4494-c2fb-415c-a38a-79afbf13be81" (UID: "9dfa4494-c2fb-415c-a38a-79afbf13be81"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.719324 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9dfa4494-c2fb-415c-a38a-79afbf13be81" (UID: "9dfa4494-c2fb-415c-a38a-79afbf13be81"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.738922 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "9dfa4494-c2fb-415c-a38a-79afbf13be81" (UID: "9dfa4494-c2fb-415c-a38a-79afbf13be81"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.784648 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbwvl\" (UniqueName: \"kubernetes.io/projected/9dfa4494-c2fb-415c-a38a-79afbf13be81-kube-api-access-qbwvl\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.784683 4829 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.784698 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.784710 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:24 crc kubenswrapper[4829]: I1002 07:57:24.784723 4829 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dfa4494-c2fb-415c-a38a-79afbf13be81-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.095897 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" event={"ID":"9dfa4494-c2fb-415c-a38a-79afbf13be81","Type":"ContainerDied","Data":"3b7c1346cfa4b6414dc016631fc30317cb921b05afcf87ca295b50eaee52d681"} Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.095968 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-94ftd" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.095990 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b7c1346cfa4b6414dc016631fc30317cb921b05afcf87ca295b50eaee52d681" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.202378 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99"] Oct 02 07:57:25 crc kubenswrapper[4829]: E1002 07:57:25.202997 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="registry-server" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.203024 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="registry-server" Oct 02 07:57:25 crc kubenswrapper[4829]: E1002 07:57:25.203040 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="extract-content" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.203051 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="extract-content" Oct 02 07:57:25 crc kubenswrapper[4829]: E1002 07:57:25.203096 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="extract-utilities" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.203105 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="extract-utilities" Oct 02 07:57:25 crc kubenswrapper[4829]: E1002 07:57:25.203113 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dfa4494-c2fb-415c-a38a-79afbf13be81" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.203124 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dfa4494-c2fb-415c-a38a-79afbf13be81" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.203366 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="05538b82-50bd-47c4-91a0-366cd3fd4f24" containerName="registry-server" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.203418 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dfa4494-c2fb-415c-a38a-79afbf13be81" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.204284 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.208802 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.208838 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.208830 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.209005 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.209176 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.209186 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.210861 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.212777 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99"] Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.294945 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lhgp\" (UniqueName: \"kubernetes.io/projected/97915d50-01b1-415e-99f1-17c7d4340c66-kube-api-access-6lhgp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295282 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295306 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295343 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295369 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295411 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295610 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/97915d50-01b1-415e-99f1-17c7d4340c66-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295706 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.295754 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.397909 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.397981 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398025 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/97915d50-01b1-415e-99f1-17c7d4340c66-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398105 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398153 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398246 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lhgp\" (UniqueName: \"kubernetes.io/projected/97915d50-01b1-415e-99f1-17c7d4340c66-kube-api-access-6lhgp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398287 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398313 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.398356 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.400939 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/97915d50-01b1-415e-99f1-17c7d4340c66-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.403415 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.405920 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.406025 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.406463 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.406972 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.407193 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.407517 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.420357 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lhgp\" (UniqueName: \"kubernetes.io/projected/97915d50-01b1-415e-99f1-17c7d4340c66-kube-api-access-6lhgp\") pod \"nova-edpm-deployment-openstack-edpm-ipam-jmb99\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:25 crc kubenswrapper[4829]: I1002 07:57:25.522342 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 07:57:26 crc kubenswrapper[4829]: I1002 07:57:26.102519 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99"] Oct 02 07:57:27 crc kubenswrapper[4829]: I1002 07:57:27.121537 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" event={"ID":"97915d50-01b1-415e-99f1-17c7d4340c66","Type":"ContainerStarted","Data":"ce7fbb41563020805747a55ad0c4bb6deecbc35cd675c1fda2424bd2add8dfef"} Oct 02 07:57:27 crc kubenswrapper[4829]: I1002 07:57:27.121971 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" event={"ID":"97915d50-01b1-415e-99f1-17c7d4340c66","Type":"ContainerStarted","Data":"44454044839e247c87e5534b3602a532666da8288601d059f7fa3e42942d21c1"} Oct 02 07:57:27 crc kubenswrapper[4829]: I1002 07:57:27.144348 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" podStartSLOduration=1.557369328 podStartE2EDuration="2.144325317s" podCreationTimestamp="2025-10-02 07:57:25 +0000 UTC" firstStartedPulling="2025-10-02 07:57:26.104593846 +0000 UTC m=+2437.444242251" lastFinishedPulling="2025-10-02 07:57:26.691549835 +0000 UTC m=+2438.031198240" observedRunningTime="2025-10-02 07:57:27.142066615 +0000 UTC m=+2438.481715060" watchObservedRunningTime="2025-10-02 07:57:27.144325317 +0000 UTC m=+2438.483973752" Oct 02 07:57:28 crc kubenswrapper[4829]: I1002 07:57:28.460636 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:57:28 crc kubenswrapper[4829]: E1002 07:57:28.460994 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:57:40 crc kubenswrapper[4829]: I1002 07:57:40.461323 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:57:40 crc kubenswrapper[4829]: E1002 07:57:40.462581 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:57:51 crc kubenswrapper[4829]: I1002 07:57:51.461491 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:57:51 crc kubenswrapper[4829]: E1002 07:57:51.462771 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:58:04 crc kubenswrapper[4829]: I1002 07:58:04.461023 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:58:04 crc kubenswrapper[4829]: E1002 07:58:04.461682 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:58:17 crc kubenswrapper[4829]: I1002 07:58:17.463126 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:58:17 crc kubenswrapper[4829]: E1002 07:58:17.464135 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.691054 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s4pnm"] Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.694802 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.709901 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4pnm"] Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.806874 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnzmh\" (UniqueName: \"kubernetes.io/projected/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-kube-api-access-qnzmh\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.807286 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-utilities\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.807496 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-catalog-content\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.909728 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-catalog-content\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.910048 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnzmh\" (UniqueName: \"kubernetes.io/projected/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-kube-api-access-qnzmh\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.910204 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-utilities\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.910341 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-catalog-content\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.910694 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-utilities\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:22 crc kubenswrapper[4829]: I1002 07:58:22.933025 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnzmh\" (UniqueName: \"kubernetes.io/projected/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-kube-api-access-qnzmh\") pod \"redhat-marketplace-s4pnm\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:23 crc kubenswrapper[4829]: I1002 07:58:23.040149 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:23 crc kubenswrapper[4829]: I1002 07:58:23.479896 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4pnm"] Oct 02 07:58:23 crc kubenswrapper[4829]: I1002 07:58:23.750404 4829 generic.go:334] "Generic (PLEG): container finished" podID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerID="2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5" exitCode=0 Oct 02 07:58:23 crc kubenswrapper[4829]: I1002 07:58:23.750450 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4pnm" event={"ID":"047df0ed-8ad6-4f11-930b-7c81c20fc0cc","Type":"ContainerDied","Data":"2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5"} Oct 02 07:58:23 crc kubenswrapper[4829]: I1002 07:58:23.750728 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4pnm" event={"ID":"047df0ed-8ad6-4f11-930b-7c81c20fc0cc","Type":"ContainerStarted","Data":"8eaefb55038e669489e853a08d05e7ef5037b662b3db76034e7d532c9ccf5e6a"} Oct 02 07:58:24 crc kubenswrapper[4829]: I1002 07:58:24.763259 4829 generic.go:334] "Generic (PLEG): container finished" podID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerID="387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939" exitCode=0 Oct 02 07:58:24 crc kubenswrapper[4829]: I1002 07:58:24.763338 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4pnm" event={"ID":"047df0ed-8ad6-4f11-930b-7c81c20fc0cc","Type":"ContainerDied","Data":"387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939"} Oct 02 07:58:25 crc kubenswrapper[4829]: I1002 07:58:25.775797 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4pnm" event={"ID":"047df0ed-8ad6-4f11-930b-7c81c20fc0cc","Type":"ContainerStarted","Data":"9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca"} Oct 02 07:58:25 crc kubenswrapper[4829]: I1002 07:58:25.807182 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s4pnm" podStartSLOduration=2.305870791 podStartE2EDuration="3.807164753s" podCreationTimestamp="2025-10-02 07:58:22 +0000 UTC" firstStartedPulling="2025-10-02 07:58:23.752276945 +0000 UTC m=+2495.091925370" lastFinishedPulling="2025-10-02 07:58:25.253570927 +0000 UTC m=+2496.593219332" observedRunningTime="2025-10-02 07:58:25.797607955 +0000 UTC m=+2497.137256360" watchObservedRunningTime="2025-10-02 07:58:25.807164753 +0000 UTC m=+2497.146813158" Oct 02 07:58:26 crc kubenswrapper[4829]: I1002 07:58:26.891625 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5w8v5"] Oct 02 07:58:26 crc kubenswrapper[4829]: I1002 07:58:26.895807 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:26 crc kubenswrapper[4829]: I1002 07:58:26.901969 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5w8v5"] Oct 02 07:58:26 crc kubenswrapper[4829]: I1002 07:58:26.990937 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-utilities\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:26 crc kubenswrapper[4829]: I1002 07:58:26.991243 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kq2z\" (UniqueName: \"kubernetes.io/projected/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-kube-api-access-9kq2z\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:26 crc kubenswrapper[4829]: I1002 07:58:26.991549 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-catalog-content\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.093293 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-catalog-content\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.093427 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-utilities\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.093522 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kq2z\" (UniqueName: \"kubernetes.io/projected/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-kube-api-access-9kq2z\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.093969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-catalog-content\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.094066 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-utilities\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.114937 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kq2z\" (UniqueName: \"kubernetes.io/projected/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-kube-api-access-9kq2z\") pod \"redhat-operators-5w8v5\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.218261 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.719102 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5w8v5"] Oct 02 07:58:27 crc kubenswrapper[4829]: I1002 07:58:27.799079 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerStarted","Data":"9a926e68589bb65332033b84ae418d24277f34da24d0b4bcec7b3c3d80017126"} Oct 02 07:58:28 crc kubenswrapper[4829]: I1002 07:58:28.811272 4829 generic.go:334] "Generic (PLEG): container finished" podID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerID="fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e" exitCode=0 Oct 02 07:58:28 crc kubenswrapper[4829]: I1002 07:58:28.811324 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerDied","Data":"fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e"} Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.087396 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5ls78"] Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.089491 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.102083 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5ls78"] Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.141932 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-catalog-content\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.142101 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bbwd\" (UniqueName: \"kubernetes.io/projected/e13d517c-bdae-4282-a1ba-9f0b76c0f363-kube-api-access-8bbwd\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.142181 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-utilities\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.244865 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-catalog-content\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.245026 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bbwd\" (UniqueName: \"kubernetes.io/projected/e13d517c-bdae-4282-a1ba-9f0b76c0f363-kube-api-access-8bbwd\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.245077 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-utilities\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.245418 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-catalog-content\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.246032 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-utilities\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.273169 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bbwd\" (UniqueName: \"kubernetes.io/projected/e13d517c-bdae-4282-a1ba-9f0b76c0f363-kube-api-access-8bbwd\") pod \"certified-operators-5ls78\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.412335 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.472729 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.821271 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"ff7d7fa7a0802a3943548b15c46ba270f460ad42261eec82e1b15ecc455b5474"} Oct 02 07:58:29 crc kubenswrapper[4829]: I1002 07:58:29.971736 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5ls78"] Oct 02 07:58:29 crc kubenswrapper[4829]: W1002 07:58:29.982974 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode13d517c_bdae_4282_a1ba_9f0b76c0f363.slice/crio-aa1c30261fc2fed9ac7a45fbb226a25b3fa10dcdf6e0d765b7bbb6502cb62de1 WatchSource:0}: Error finding container aa1c30261fc2fed9ac7a45fbb226a25b3fa10dcdf6e0d765b7bbb6502cb62de1: Status 404 returned error can't find the container with id aa1c30261fc2fed9ac7a45fbb226a25b3fa10dcdf6e0d765b7bbb6502cb62de1 Oct 02 07:58:30 crc kubenswrapper[4829]: I1002 07:58:30.835811 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerStarted","Data":"ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c"} Oct 02 07:58:30 crc kubenswrapper[4829]: I1002 07:58:30.838299 4829 generic.go:334] "Generic (PLEG): container finished" podID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerID="380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b" exitCode=0 Oct 02 07:58:30 crc kubenswrapper[4829]: I1002 07:58:30.838416 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerDied","Data":"380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b"} Oct 02 07:58:30 crc kubenswrapper[4829]: I1002 07:58:30.838509 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerStarted","Data":"aa1c30261fc2fed9ac7a45fbb226a25b3fa10dcdf6e0d765b7bbb6502cb62de1"} Oct 02 07:58:31 crc kubenswrapper[4829]: I1002 07:58:31.854633 4829 generic.go:334] "Generic (PLEG): container finished" podID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerID="ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c" exitCode=0 Oct 02 07:58:31 crc kubenswrapper[4829]: I1002 07:58:31.854739 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerDied","Data":"ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c"} Oct 02 07:58:32 crc kubenswrapper[4829]: I1002 07:58:32.869345 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerStarted","Data":"acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358"} Oct 02 07:58:32 crc kubenswrapper[4829]: I1002 07:58:32.872824 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerStarted","Data":"519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6"} Oct 02 07:58:32 crc kubenswrapper[4829]: I1002 07:58:32.886730 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5w8v5" podStartSLOduration=3.373404196 podStartE2EDuration="6.88670783s" podCreationTimestamp="2025-10-02 07:58:26 +0000 UTC" firstStartedPulling="2025-10-02 07:58:28.81465251 +0000 UTC m=+2500.154300955" lastFinishedPulling="2025-10-02 07:58:32.327956174 +0000 UTC m=+2503.667604589" observedRunningTime="2025-10-02 07:58:32.886528325 +0000 UTC m=+2504.226176740" watchObservedRunningTime="2025-10-02 07:58:32.88670783 +0000 UTC m=+2504.226356235" Oct 02 07:58:33 crc kubenswrapper[4829]: I1002 07:58:33.043886 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:33 crc kubenswrapper[4829]: I1002 07:58:33.044278 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:33 crc kubenswrapper[4829]: I1002 07:58:33.101181 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:33 crc kubenswrapper[4829]: I1002 07:58:33.882696 4829 generic.go:334] "Generic (PLEG): container finished" podID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerID="519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6" exitCode=0 Oct 02 07:58:33 crc kubenswrapper[4829]: I1002 07:58:33.882853 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerDied","Data":"519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6"} Oct 02 07:58:33 crc kubenswrapper[4829]: I1002 07:58:33.947657 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:34 crc kubenswrapper[4829]: I1002 07:58:34.896432 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerStarted","Data":"c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7"} Oct 02 07:58:34 crc kubenswrapper[4829]: I1002 07:58:34.925873 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5ls78" podStartSLOduration=2.380977369 podStartE2EDuration="5.925847476s" podCreationTimestamp="2025-10-02 07:58:29 +0000 UTC" firstStartedPulling="2025-10-02 07:58:30.84168858 +0000 UTC m=+2502.181337015" lastFinishedPulling="2025-10-02 07:58:34.386558717 +0000 UTC m=+2505.726207122" observedRunningTime="2025-10-02 07:58:34.916496256 +0000 UTC m=+2506.256144671" watchObservedRunningTime="2025-10-02 07:58:34.925847476 +0000 UTC m=+2506.265495891" Oct 02 07:58:36 crc kubenswrapper[4829]: I1002 07:58:36.882650 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4pnm"] Oct 02 07:58:36 crc kubenswrapper[4829]: I1002 07:58:36.917370 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s4pnm" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="registry-server" containerID="cri-o://9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca" gracePeriod=2 Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.218454 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.218778 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.469566 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.525893 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-utilities\") pod \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.526134 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnzmh\" (UniqueName: \"kubernetes.io/projected/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-kube-api-access-qnzmh\") pod \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.526216 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-catalog-content\") pod \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\" (UID: \"047df0ed-8ad6-4f11-930b-7c81c20fc0cc\") " Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.526620 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-utilities" (OuterVolumeSpecName: "utilities") pod "047df0ed-8ad6-4f11-930b-7c81c20fc0cc" (UID: "047df0ed-8ad6-4f11-930b-7c81c20fc0cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.526782 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.535909 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-kube-api-access-qnzmh" (OuterVolumeSpecName: "kube-api-access-qnzmh") pod "047df0ed-8ad6-4f11-930b-7c81c20fc0cc" (UID: "047df0ed-8ad6-4f11-930b-7c81c20fc0cc"). InnerVolumeSpecName "kube-api-access-qnzmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.546500 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "047df0ed-8ad6-4f11-930b-7c81c20fc0cc" (UID: "047df0ed-8ad6-4f11-930b-7c81c20fc0cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.629176 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnzmh\" (UniqueName: \"kubernetes.io/projected/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-kube-api-access-qnzmh\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.629638 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/047df0ed-8ad6-4f11-930b-7c81c20fc0cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.931058 4829 generic.go:334] "Generic (PLEG): container finished" podID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerID="9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca" exitCode=0 Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.931135 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4pnm" event={"ID":"047df0ed-8ad6-4f11-930b-7c81c20fc0cc","Type":"ContainerDied","Data":"9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca"} Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.931282 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s4pnm" event={"ID":"047df0ed-8ad6-4f11-930b-7c81c20fc0cc","Type":"ContainerDied","Data":"8eaefb55038e669489e853a08d05e7ef5037b662b3db76034e7d532c9ccf5e6a"} Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.931321 4829 scope.go:117] "RemoveContainer" containerID="9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.931166 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s4pnm" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.963982 4829 scope.go:117] "RemoveContainer" containerID="387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939" Oct 02 07:58:37 crc kubenswrapper[4829]: I1002 07:58:37.991326 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4pnm"] Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.001329 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s4pnm"] Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.010109 4829 scope.go:117] "RemoveContainer" containerID="2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.061574 4829 scope.go:117] "RemoveContainer" containerID="9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca" Oct 02 07:58:38 crc kubenswrapper[4829]: E1002 07:58:38.062348 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca\": container with ID starting with 9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca not found: ID does not exist" containerID="9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.062635 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca"} err="failed to get container status \"9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca\": rpc error: code = NotFound desc = could not find container \"9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca\": container with ID starting with 9b01c690b54ec1e7ddd737ed4ac6dde5b45021d379ef69adcc73b12270e911ca not found: ID does not exist" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.062703 4829 scope.go:117] "RemoveContainer" containerID="387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939" Oct 02 07:58:38 crc kubenswrapper[4829]: E1002 07:58:38.063423 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939\": container with ID starting with 387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939 not found: ID does not exist" containerID="387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.063498 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939"} err="failed to get container status \"387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939\": rpc error: code = NotFound desc = could not find container \"387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939\": container with ID starting with 387a507be48dac4c541a264e7c1ef6b8993bc7879c693ef7678ddc567e5d2939 not found: ID does not exist" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.063556 4829 scope.go:117] "RemoveContainer" containerID="2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5" Oct 02 07:58:38 crc kubenswrapper[4829]: E1002 07:58:38.064576 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5\": container with ID starting with 2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5 not found: ID does not exist" containerID="2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.064640 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5"} err="failed to get container status \"2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5\": rpc error: code = NotFound desc = could not find container \"2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5\": container with ID starting with 2fdfa9c181bc43702b54239e4658df04dfeb511a4007a1afb0ba721e772a2ce5 not found: ID does not exist" Oct 02 07:58:38 crc kubenswrapper[4829]: I1002 07:58:38.276960 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5w8v5" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="registry-server" probeResult="failure" output=< Oct 02 07:58:38 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 07:58:38 crc kubenswrapper[4829]: > Oct 02 07:58:39 crc kubenswrapper[4829]: I1002 07:58:39.413585 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:39 crc kubenswrapper[4829]: I1002 07:58:39.413886 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:39 crc kubenswrapper[4829]: I1002 07:58:39.481699 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" path="/var/lib/kubelet/pods/047df0ed-8ad6-4f11-930b-7c81c20fc0cc/volumes" Oct 02 07:58:39 crc kubenswrapper[4829]: I1002 07:58:39.501350 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:40 crc kubenswrapper[4829]: I1002 07:58:40.001621 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:40 crc kubenswrapper[4829]: I1002 07:58:40.484493 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5ls78"] Oct 02 07:58:41 crc kubenswrapper[4829]: I1002 07:58:41.977157 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5ls78" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="registry-server" containerID="cri-o://c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7" gracePeriod=2 Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.475728 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.532084 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-catalog-content\") pod \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.532195 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-utilities\") pod \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.533218 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-utilities" (OuterVolumeSpecName: "utilities") pod "e13d517c-bdae-4282-a1ba-9f0b76c0f363" (UID: "e13d517c-bdae-4282-a1ba-9f0b76c0f363"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.533539 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bbwd\" (UniqueName: \"kubernetes.io/projected/e13d517c-bdae-4282-a1ba-9f0b76c0f363-kube-api-access-8bbwd\") pod \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\" (UID: \"e13d517c-bdae-4282-a1ba-9f0b76c0f363\") " Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.534876 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.539924 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13d517c-bdae-4282-a1ba-9f0b76c0f363-kube-api-access-8bbwd" (OuterVolumeSpecName: "kube-api-access-8bbwd") pod "e13d517c-bdae-4282-a1ba-9f0b76c0f363" (UID: "e13d517c-bdae-4282-a1ba-9f0b76c0f363"). InnerVolumeSpecName "kube-api-access-8bbwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.605286 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e13d517c-bdae-4282-a1ba-9f0b76c0f363" (UID: "e13d517c-bdae-4282-a1ba-9f0b76c0f363"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.636809 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bbwd\" (UniqueName: \"kubernetes.io/projected/e13d517c-bdae-4282-a1ba-9f0b76c0f363-kube-api-access-8bbwd\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.636840 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e13d517c-bdae-4282-a1ba-9f0b76c0f363-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.995653 4829 generic.go:334] "Generic (PLEG): container finished" podID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerID="c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7" exitCode=0 Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.995727 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerDied","Data":"c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7"} Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.995771 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5ls78" Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.995821 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ls78" event={"ID":"e13d517c-bdae-4282-a1ba-9f0b76c0f363","Type":"ContainerDied","Data":"aa1c30261fc2fed9ac7a45fbb226a25b3fa10dcdf6e0d765b7bbb6502cb62de1"} Oct 02 07:58:42 crc kubenswrapper[4829]: I1002 07:58:42.995856 4829 scope.go:117] "RemoveContainer" containerID="c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.023636 4829 scope.go:117] "RemoveContainer" containerID="519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.058009 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5ls78"] Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.077019 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5ls78"] Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.090219 4829 scope.go:117] "RemoveContainer" containerID="380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.132103 4829 scope.go:117] "RemoveContainer" containerID="c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7" Oct 02 07:58:43 crc kubenswrapper[4829]: E1002 07:58:43.132594 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7\": container with ID starting with c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7 not found: ID does not exist" containerID="c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.132637 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7"} err="failed to get container status \"c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7\": rpc error: code = NotFound desc = could not find container \"c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7\": container with ID starting with c0cbcfb96a3fb00ecf895a8bd58052aaee81826b501891754c37e89030cfd3a7 not found: ID does not exist" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.132666 4829 scope.go:117] "RemoveContainer" containerID="519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6" Oct 02 07:58:43 crc kubenswrapper[4829]: E1002 07:58:43.133546 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6\": container with ID starting with 519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6 not found: ID does not exist" containerID="519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.133589 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6"} err="failed to get container status \"519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6\": rpc error: code = NotFound desc = could not find container \"519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6\": container with ID starting with 519e12c0b54938d389e44981d8990153c8d41def5f07346be017d2d1165f4ab6 not found: ID does not exist" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.133606 4829 scope.go:117] "RemoveContainer" containerID="380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b" Oct 02 07:58:43 crc kubenswrapper[4829]: E1002 07:58:43.133943 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b\": container with ID starting with 380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b not found: ID does not exist" containerID="380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.133985 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b"} err="failed to get container status \"380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b\": rpc error: code = NotFound desc = could not find container \"380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b\": container with ID starting with 380eaba4fb97e8b113f22820a96a388b9b33e5c618a92c70c9e90af1f21ef52b not found: ID does not exist" Oct 02 07:58:43 crc kubenswrapper[4829]: I1002 07:58:43.481758 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" path="/var/lib/kubelet/pods/e13d517c-bdae-4282-a1ba-9f0b76c0f363/volumes" Oct 02 07:58:47 crc kubenswrapper[4829]: I1002 07:58:47.287175 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:47 crc kubenswrapper[4829]: I1002 07:58:47.339194 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:47 crc kubenswrapper[4829]: I1002 07:58:47.524944 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5w8v5"] Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.070655 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5w8v5" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="registry-server" containerID="cri-o://acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358" gracePeriod=2 Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.565431 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.700238 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-catalog-content\") pod \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.700680 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kq2z\" (UniqueName: \"kubernetes.io/projected/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-kube-api-access-9kq2z\") pod \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.700740 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-utilities\") pod \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\" (UID: \"c6c1c768-ea4c-4dd5-94e3-d50f449fce76\") " Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.701823 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-utilities" (OuterVolumeSpecName: "utilities") pod "c6c1c768-ea4c-4dd5-94e3-d50f449fce76" (UID: "c6c1c768-ea4c-4dd5-94e3-d50f449fce76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.708123 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-kube-api-access-9kq2z" (OuterVolumeSpecName: "kube-api-access-9kq2z") pod "c6c1c768-ea4c-4dd5-94e3-d50f449fce76" (UID: "c6c1c768-ea4c-4dd5-94e3-d50f449fce76"). InnerVolumeSpecName "kube-api-access-9kq2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.784031 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c6c1c768-ea4c-4dd5-94e3-d50f449fce76" (UID: "c6c1c768-ea4c-4dd5-94e3-d50f449fce76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.803098 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kq2z\" (UniqueName: \"kubernetes.io/projected/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-kube-api-access-9kq2z\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.803130 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:49 crc kubenswrapper[4829]: I1002 07:58:49.803139 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c6c1c768-ea4c-4dd5-94e3-d50f449fce76-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.089710 4829 generic.go:334] "Generic (PLEG): container finished" podID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerID="acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358" exitCode=0 Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.089795 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerDied","Data":"acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358"} Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.089849 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w8v5" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.089884 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w8v5" event={"ID":"c6c1c768-ea4c-4dd5-94e3-d50f449fce76","Type":"ContainerDied","Data":"9a926e68589bb65332033b84ae418d24277f34da24d0b4bcec7b3c3d80017126"} Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.089919 4829 scope.go:117] "RemoveContainer" containerID="acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.116312 4829 scope.go:117] "RemoveContainer" containerID="ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.150785 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5w8v5"] Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.160763 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5w8v5"] Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.164205 4829 scope.go:117] "RemoveContainer" containerID="fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.232516 4829 scope.go:117] "RemoveContainer" containerID="acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358" Oct 02 07:58:50 crc kubenswrapper[4829]: E1002 07:58:50.233282 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358\": container with ID starting with acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358 not found: ID does not exist" containerID="acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.233771 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358"} err="failed to get container status \"acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358\": rpc error: code = NotFound desc = could not find container \"acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358\": container with ID starting with acc50dd50288a401ce69f8dddf4d97d6a58cf3009494dea116bd7f2614dc4358 not found: ID does not exist" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.233898 4829 scope.go:117] "RemoveContainer" containerID="ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c" Oct 02 07:58:50 crc kubenswrapper[4829]: E1002 07:58:50.234499 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c\": container with ID starting with ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c not found: ID does not exist" containerID="ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.234528 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c"} err="failed to get container status \"ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c\": rpc error: code = NotFound desc = could not find container \"ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c\": container with ID starting with ff0617a15cd4194cb0c1ce6fdf05a574adad04306a4331ccb2f1d7953ec7da6c not found: ID does not exist" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.234546 4829 scope.go:117] "RemoveContainer" containerID="fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e" Oct 02 07:58:50 crc kubenswrapper[4829]: E1002 07:58:50.235152 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e\": container with ID starting with fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e not found: ID does not exist" containerID="fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e" Oct 02 07:58:50 crc kubenswrapper[4829]: I1002 07:58:50.235203 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e"} err="failed to get container status \"fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e\": rpc error: code = NotFound desc = could not find container \"fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e\": container with ID starting with fd772b9a6e31cd7e5b004880b1807a4d97d82ab2b36be42b5e8599068eade63e not found: ID does not exist" Oct 02 07:58:51 crc kubenswrapper[4829]: I1002 07:58:51.482691 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" path="/var/lib/kubelet/pods/c6c1c768-ea4c-4dd5-94e3-d50f449fce76/volumes" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.153009 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl"] Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.153958 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="extract-utilities" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.153975 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="extract-utilities" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154009 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="extract-utilities" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154017 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="extract-utilities" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154032 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="extract-content" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154041 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="extract-content" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154058 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154068 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154083 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154092 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154107 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="extract-utilities" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154115 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="extract-utilities" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154134 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154144 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154156 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="extract-content" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154165 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="extract-content" Oct 02 08:00:00 crc kubenswrapper[4829]: E1002 08:00:00.154188 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="extract-content" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154197 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="extract-content" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154468 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6c1c768-ea4c-4dd5-94e3-d50f449fce76" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154488 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="047df0ed-8ad6-4f11-930b-7c81c20fc0cc" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.154503 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13d517c-bdae-4282-a1ba-9f0b76c0f363" containerName="registry-server" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.155318 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.157943 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79t2r\" (UniqueName: \"kubernetes.io/projected/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-kube-api-access-79t2r\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.158056 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-config-volume\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.158090 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-secret-volume\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.159165 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.161055 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.171338 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl"] Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.259833 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-config-volume\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.259913 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-secret-volume\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.260044 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79t2r\" (UniqueName: \"kubernetes.io/projected/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-kube-api-access-79t2r\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.260882 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-config-volume\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.267496 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-secret-volume\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.282739 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79t2r\" (UniqueName: \"kubernetes.io/projected/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-kube-api-access-79t2r\") pod \"collect-profiles-29323200-nqwsl\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.479578 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:00 crc kubenswrapper[4829]: I1002 08:00:00.953578 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl"] Oct 02 08:00:01 crc kubenswrapper[4829]: I1002 08:00:01.937622 4829 generic.go:334] "Generic (PLEG): container finished" podID="4e33017a-e2ee-470c-b9e5-03b4602ff7a0" containerID="a705f53b9287a8c09aedb390e28d761082a3ca0779d24d5fc09c4fe742410d32" exitCode=0 Oct 02 08:00:01 crc kubenswrapper[4829]: I1002 08:00:01.937671 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" event={"ID":"4e33017a-e2ee-470c-b9e5-03b4602ff7a0","Type":"ContainerDied","Data":"a705f53b9287a8c09aedb390e28d761082a3ca0779d24d5fc09c4fe742410d32"} Oct 02 08:00:01 crc kubenswrapper[4829]: I1002 08:00:01.937960 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" event={"ID":"4e33017a-e2ee-470c-b9e5-03b4602ff7a0","Type":"ContainerStarted","Data":"fbe3759c7a9ce62ae4da2a8de46ed34d05b34e0bdd7094f87063539eea6edd0b"} Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.406422 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.429417 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-config-volume\") pod \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.429507 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79t2r\" (UniqueName: \"kubernetes.io/projected/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-kube-api-access-79t2r\") pod \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.429578 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-secret-volume\") pod \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\" (UID: \"4e33017a-e2ee-470c-b9e5-03b4602ff7a0\") " Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.430299 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-config-volume" (OuterVolumeSpecName: "config-volume") pod "4e33017a-e2ee-470c-b9e5-03b4602ff7a0" (UID: "4e33017a-e2ee-470c-b9e5-03b4602ff7a0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.430668 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.453873 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4e33017a-e2ee-470c-b9e5-03b4602ff7a0" (UID: "4e33017a-e2ee-470c-b9e5-03b4602ff7a0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.456051 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-kube-api-access-79t2r" (OuterVolumeSpecName: "kube-api-access-79t2r") pod "4e33017a-e2ee-470c-b9e5-03b4602ff7a0" (UID: "4e33017a-e2ee-470c-b9e5-03b4602ff7a0"). InnerVolumeSpecName "kube-api-access-79t2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.531749 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79t2r\" (UniqueName: \"kubernetes.io/projected/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-kube-api-access-79t2r\") on node \"crc\" DevicePath \"\"" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.531796 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4e33017a-e2ee-470c-b9e5-03b4602ff7a0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.960818 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" event={"ID":"4e33017a-e2ee-470c-b9e5-03b4602ff7a0","Type":"ContainerDied","Data":"fbe3759c7a9ce62ae4da2a8de46ed34d05b34e0bdd7094f87063539eea6edd0b"} Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.960864 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fbe3759c7a9ce62ae4da2a8de46ed34d05b34e0bdd7094f87063539eea6edd0b" Oct 02 08:00:03 crc kubenswrapper[4829]: I1002 08:00:03.960869 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl" Oct 02 08:00:04 crc kubenswrapper[4829]: I1002 08:00:04.480889 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs"] Oct 02 08:00:04 crc kubenswrapper[4829]: I1002 08:00:04.488743 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323155-g97qs"] Oct 02 08:00:05 crc kubenswrapper[4829]: I1002 08:00:05.479993 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21f59a4c-e5a4-4b81-8c74-82d7cd1db54a" path="/var/lib/kubelet/pods/21f59a4c-e5a4-4b81-8c74-82d7cd1db54a/volumes" Oct 02 08:00:55 crc kubenswrapper[4829]: I1002 08:00:55.329655 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:00:55 crc kubenswrapper[4829]: I1002 08:00:55.330422 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:00:59 crc kubenswrapper[4829]: I1002 08:00:59.234020 4829 scope.go:117] "RemoveContainer" containerID="30c219a07ea7a3e8f0dd4125c1dd8f2a58047f3194b0c451a53d8875cd1eee71" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.159378 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323201-j6v6q"] Oct 02 08:01:00 crc kubenswrapper[4829]: E1002 08:01:00.159975 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e33017a-e2ee-470c-b9e5-03b4602ff7a0" containerName="collect-profiles" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.159986 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e33017a-e2ee-470c-b9e5-03b4602ff7a0" containerName="collect-profiles" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.160159 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e33017a-e2ee-470c-b9e5-03b4602ff7a0" containerName="collect-profiles" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.160777 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.203552 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323201-j6v6q"] Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.228768 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-config-data\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.228905 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-fernet-keys\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.229032 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-combined-ca-bundle\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.229063 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvtxc\" (UniqueName: \"kubernetes.io/projected/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-kube-api-access-wvtxc\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.330773 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-fernet-keys\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.332045 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-combined-ca-bundle\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.332345 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvtxc\" (UniqueName: \"kubernetes.io/projected/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-kube-api-access-wvtxc\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.332745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-config-data\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.336942 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-fernet-keys\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.340084 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-config-data\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.340158 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-combined-ca-bundle\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.362145 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvtxc\" (UniqueName: \"kubernetes.io/projected/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-kube-api-access-wvtxc\") pod \"keystone-cron-29323201-j6v6q\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.486766 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:00 crc kubenswrapper[4829]: I1002 08:01:00.950335 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323201-j6v6q"] Oct 02 08:01:01 crc kubenswrapper[4829]: I1002 08:01:01.680894 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-j6v6q" event={"ID":"0ca1b8dd-db65-4bcf-908c-3bf2b145d334","Type":"ContainerStarted","Data":"2e1edf1fc0cac367a7ad4b2af4523c4552cba0b6002b5e74b62f8a2548e2dbb9"} Oct 02 08:01:01 crc kubenswrapper[4829]: I1002 08:01:01.681264 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-j6v6q" event={"ID":"0ca1b8dd-db65-4bcf-908c-3bf2b145d334","Type":"ContainerStarted","Data":"ae0c4f9083afc166a5ba3be27f8bb414b46ad283924737fef847e56faeda3ca9"} Oct 02 08:01:01 crc kubenswrapper[4829]: I1002 08:01:01.707906 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323201-j6v6q" podStartSLOduration=1.70788896 podStartE2EDuration="1.70788896s" podCreationTimestamp="2025-10-02 08:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:01:01.704818024 +0000 UTC m=+2653.044466449" watchObservedRunningTime="2025-10-02 08:01:01.70788896 +0000 UTC m=+2653.047537375" Oct 02 08:01:03 crc kubenswrapper[4829]: I1002 08:01:03.709170 4829 generic.go:334] "Generic (PLEG): container finished" podID="0ca1b8dd-db65-4bcf-908c-3bf2b145d334" containerID="2e1edf1fc0cac367a7ad4b2af4523c4552cba0b6002b5e74b62f8a2548e2dbb9" exitCode=0 Oct 02 08:01:03 crc kubenswrapper[4829]: I1002 08:01:03.709307 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-j6v6q" event={"ID":"0ca1b8dd-db65-4bcf-908c-3bf2b145d334","Type":"ContainerDied","Data":"2e1edf1fc0cac367a7ad4b2af4523c4552cba0b6002b5e74b62f8a2548e2dbb9"} Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.090637 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.135121 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-config-data\") pod \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.135373 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvtxc\" (UniqueName: \"kubernetes.io/projected/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-kube-api-access-wvtxc\") pod \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.135463 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-combined-ca-bundle\") pod \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.135519 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-fernet-keys\") pod \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\" (UID: \"0ca1b8dd-db65-4bcf-908c-3bf2b145d334\") " Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.147418 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "0ca1b8dd-db65-4bcf-908c-3bf2b145d334" (UID: "0ca1b8dd-db65-4bcf-908c-3bf2b145d334"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.150643 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-kube-api-access-wvtxc" (OuterVolumeSpecName: "kube-api-access-wvtxc") pod "0ca1b8dd-db65-4bcf-908c-3bf2b145d334" (UID: "0ca1b8dd-db65-4bcf-908c-3bf2b145d334"). InnerVolumeSpecName "kube-api-access-wvtxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.169786 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ca1b8dd-db65-4bcf-908c-3bf2b145d334" (UID: "0ca1b8dd-db65-4bcf-908c-3bf2b145d334"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.201561 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-config-data" (OuterVolumeSpecName: "config-data") pod "0ca1b8dd-db65-4bcf-908c-3bf2b145d334" (UID: "0ca1b8dd-db65-4bcf-908c-3bf2b145d334"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.236757 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.236794 4829 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.236807 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.236819 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvtxc\" (UniqueName: \"kubernetes.io/projected/0ca1b8dd-db65-4bcf-908c-3bf2b145d334-kube-api-access-wvtxc\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.735218 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323201-j6v6q" event={"ID":"0ca1b8dd-db65-4bcf-908c-3bf2b145d334","Type":"ContainerDied","Data":"ae0c4f9083afc166a5ba3be27f8bb414b46ad283924737fef847e56faeda3ca9"} Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.735688 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae0c4f9083afc166a5ba3be27f8bb414b46ad283924737fef847e56faeda3ca9" Oct 02 08:01:05 crc kubenswrapper[4829]: I1002 08:01:05.735552 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323201-j6v6q" Oct 02 08:01:14 crc kubenswrapper[4829]: I1002 08:01:14.864831 4829 generic.go:334] "Generic (PLEG): container finished" podID="97915d50-01b1-415e-99f1-17c7d4340c66" containerID="ce7fbb41563020805747a55ad0c4bb6deecbc35cd675c1fda2424bd2add8dfef" exitCode=0 Oct 02 08:01:14 crc kubenswrapper[4829]: I1002 08:01:14.864947 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" event={"ID":"97915d50-01b1-415e-99f1-17c7d4340c66","Type":"ContainerDied","Data":"ce7fbb41563020805747a55ad0c4bb6deecbc35cd675c1fda2424bd2add8dfef"} Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.392442 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514395 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-inventory\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514478 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-1\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514536 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-combined-ca-bundle\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514573 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lhgp\" (UniqueName: \"kubernetes.io/projected/97915d50-01b1-415e-99f1-17c7d4340c66-kube-api-access-6lhgp\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514617 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/97915d50-01b1-415e-99f1-17c7d4340c66-nova-extra-config-0\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514649 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-0\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514750 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-1\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514829 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-ssh-key\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.514955 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-0\") pod \"97915d50-01b1-415e-99f1-17c7d4340c66\" (UID: \"97915d50-01b1-415e-99f1-17c7d4340c66\") " Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.522334 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97915d50-01b1-415e-99f1-17c7d4340c66-kube-api-access-6lhgp" (OuterVolumeSpecName: "kube-api-access-6lhgp") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "kube-api-access-6lhgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.546108 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.551694 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.553874 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.554822 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.560171 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-inventory" (OuterVolumeSpecName: "inventory") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.562421 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.581287 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97915d50-01b1-415e-99f1-17c7d4340c66-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.584843 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "97915d50-01b1-415e-99f1-17c7d4340c66" (UID: "97915d50-01b1-415e-99f1-17c7d4340c66"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618747 4829 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618777 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618788 4829 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618796 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618804 4829 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618814 4829 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618822 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lhgp\" (UniqueName: \"kubernetes.io/projected/97915d50-01b1-415e-99f1-17c7d4340c66-kube-api-access-6lhgp\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618830 4829 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/97915d50-01b1-415e-99f1-17c7d4340c66-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.618839 4829 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/97915d50-01b1-415e-99f1-17c7d4340c66-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.886087 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" event={"ID":"97915d50-01b1-415e-99f1-17c7d4340c66","Type":"ContainerDied","Data":"44454044839e247c87e5534b3602a532666da8288601d059f7fa3e42942d21c1"} Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.886141 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44454044839e247c87e5534b3602a532666da8288601d059f7fa3e42942d21c1" Oct 02 08:01:16 crc kubenswrapper[4829]: I1002 08:01:16.886142 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-jmb99" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.043492 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx"] Oct 02 08:01:17 crc kubenswrapper[4829]: E1002 08:01:17.043899 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97915d50-01b1-415e-99f1-17c7d4340c66" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.043922 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="97915d50-01b1-415e-99f1-17c7d4340c66" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 08:01:17 crc kubenswrapper[4829]: E1002 08:01:17.043934 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca1b8dd-db65-4bcf-908c-3bf2b145d334" containerName="keystone-cron" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.043940 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca1b8dd-db65-4bcf-908c-3bf2b145d334" containerName="keystone-cron" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.044167 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="97915d50-01b1-415e-99f1-17c7d4340c66" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.044182 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca1b8dd-db65-4bcf-908c-3bf2b145d334" containerName="keystone-cron" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.044913 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.047154 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.047740 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.048650 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.048830 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.049004 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-kfv8q" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.072997 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx"] Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229008 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229075 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229094 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229143 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229171 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq6rk\" (UniqueName: \"kubernetes.io/projected/e74d0c61-084d-49d1-877a-3325875d5c15-kube-api-access-cq6rk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229196 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.229252 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.331002 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.331511 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq6rk\" (UniqueName: \"kubernetes.io/projected/e74d0c61-084d-49d1-877a-3325875d5c15-kube-api-access-cq6rk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.331823 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.332174 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.332627 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.332921 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.333210 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.338638 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.339004 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.339148 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.340113 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.340730 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.341113 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.353371 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq6rk\" (UniqueName: \"kubernetes.io/projected/e74d0c61-084d-49d1-877a-3325875d5c15-kube-api-access-cq6rk\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.369779 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.965494 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx"] Oct 02 08:01:17 crc kubenswrapper[4829]: W1002 08:01:17.992815 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode74d0c61_084d_49d1_877a_3325875d5c15.slice/crio-59101b465c791e3716f174eb401d8720e574476927435a56dc34439abda3fda2 WatchSource:0}: Error finding container 59101b465c791e3716f174eb401d8720e574476927435a56dc34439abda3fda2: Status 404 returned error can't find the container with id 59101b465c791e3716f174eb401d8720e574476927435a56dc34439abda3fda2 Oct 02 08:01:17 crc kubenswrapper[4829]: I1002 08:01:17.996878 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:01:18 crc kubenswrapper[4829]: I1002 08:01:18.918918 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" event={"ID":"e74d0c61-084d-49d1-877a-3325875d5c15","Type":"ContainerStarted","Data":"87c8a86d2a7078a877e1cd6e46829ff410f62e2725fe59b929654e44aa24153b"} Oct 02 08:01:18 crc kubenswrapper[4829]: I1002 08:01:18.919181 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" event={"ID":"e74d0c61-084d-49d1-877a-3325875d5c15","Type":"ContainerStarted","Data":"59101b465c791e3716f174eb401d8720e574476927435a56dc34439abda3fda2"} Oct 02 08:01:18 crc kubenswrapper[4829]: I1002 08:01:18.945500 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" podStartSLOduration=1.507549737 podStartE2EDuration="1.945475062s" podCreationTimestamp="2025-10-02 08:01:17 +0000 UTC" firstStartedPulling="2025-10-02 08:01:17.996657101 +0000 UTC m=+2669.336305496" lastFinishedPulling="2025-10-02 08:01:18.434582406 +0000 UTC m=+2669.774230821" observedRunningTime="2025-10-02 08:01:18.940622321 +0000 UTC m=+2670.280270746" watchObservedRunningTime="2025-10-02 08:01:18.945475062 +0000 UTC m=+2670.285123477" Oct 02 08:01:25 crc kubenswrapper[4829]: I1002 08:01:25.329454 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:01:25 crc kubenswrapper[4829]: I1002 08:01:25.330111 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:01:55 crc kubenswrapper[4829]: I1002 08:01:55.328956 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:01:55 crc kubenswrapper[4829]: I1002 08:01:55.329494 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:01:55 crc kubenswrapper[4829]: I1002 08:01:55.329537 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:01:55 crc kubenswrapper[4829]: I1002 08:01:55.330239 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff7d7fa7a0802a3943548b15c46ba270f460ad42261eec82e1b15ecc455b5474"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:01:55 crc kubenswrapper[4829]: I1002 08:01:55.330293 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://ff7d7fa7a0802a3943548b15c46ba270f460ad42261eec82e1b15ecc455b5474" gracePeriod=600 Oct 02 08:01:56 crc kubenswrapper[4829]: I1002 08:01:56.380569 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="ff7d7fa7a0802a3943548b15c46ba270f460ad42261eec82e1b15ecc455b5474" exitCode=0 Oct 02 08:01:56 crc kubenswrapper[4829]: I1002 08:01:56.380592 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"ff7d7fa7a0802a3943548b15c46ba270f460ad42261eec82e1b15ecc455b5474"} Oct 02 08:01:56 crc kubenswrapper[4829]: I1002 08:01:56.381152 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce"} Oct 02 08:01:56 crc kubenswrapper[4829]: I1002 08:01:56.381176 4829 scope.go:117] "RemoveContainer" containerID="1b7f77ac7a0acb6da3577f6c052262ce5ea80ad03ddbcbaf08484fd1f18d58e4" Oct 02 08:03:55 crc kubenswrapper[4829]: I1002 08:03:55.329645 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:03:55 crc kubenswrapper[4829]: I1002 08:03:55.330493 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:03:58 crc kubenswrapper[4829]: I1002 08:03:58.914379 4829 generic.go:334] "Generic (PLEG): container finished" podID="e74d0c61-084d-49d1-877a-3325875d5c15" containerID="87c8a86d2a7078a877e1cd6e46829ff410f62e2725fe59b929654e44aa24153b" exitCode=0 Oct 02 08:03:58 crc kubenswrapper[4829]: I1002 08:03:58.914505 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" event={"ID":"e74d0c61-084d-49d1-877a-3325875d5c15","Type":"ContainerDied","Data":"87c8a86d2a7078a877e1cd6e46829ff410f62e2725fe59b929654e44aa24153b"} Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.505723 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.654875 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-1\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.655024 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ssh-key\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.655197 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-telemetry-combined-ca-bundle\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.655285 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-0\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.655496 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-2\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.655590 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq6rk\" (UniqueName: \"kubernetes.io/projected/e74d0c61-084d-49d1-877a-3325875d5c15-kube-api-access-cq6rk\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.655635 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-inventory\") pod \"e74d0c61-084d-49d1-877a-3325875d5c15\" (UID: \"e74d0c61-084d-49d1-877a-3325875d5c15\") " Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.680423 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.680539 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74d0c61-084d-49d1-877a-3325875d5c15-kube-api-access-cq6rk" (OuterVolumeSpecName: "kube-api-access-cq6rk") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "kube-api-access-cq6rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.685116 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.687602 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-inventory" (OuterVolumeSpecName: "inventory") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.690373 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.691425 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.698479 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "e74d0c61-084d-49d1-877a-3325875d5c15" (UID: "e74d0c61-084d-49d1-877a-3325875d5c15"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758800 4829 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758843 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758856 4829 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758869 4829 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758880 4829 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758893 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq6rk\" (UniqueName: \"kubernetes.io/projected/e74d0c61-084d-49d1-877a-3325875d5c15-kube-api-access-cq6rk\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.758907 4829 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e74d0c61-084d-49d1-877a-3325875d5c15-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.950609 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" event={"ID":"e74d0c61-084d-49d1-877a-3325875d5c15","Type":"ContainerDied","Data":"59101b465c791e3716f174eb401d8720e574476927435a56dc34439abda3fda2"} Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.950688 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59101b465c791e3716f174eb401d8720e574476927435a56dc34439abda3fda2" Oct 02 08:04:00 crc kubenswrapper[4829]: I1002 08:04:00.950844 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx" Oct 02 08:04:03 crc kubenswrapper[4829]: E1002 08:04:03.077685 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Oct 02 08:04:25 crc kubenswrapper[4829]: I1002 08:04:25.329131 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:04:25 crc kubenswrapper[4829]: I1002 08:04:25.329876 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.090388 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.091904 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="prometheus" containerID="cri-o://b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407" gracePeriod=600 Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.092011 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="thanos-sidecar" containerID="cri-o://83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a" gracePeriod=600 Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.092413 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="config-reloader" containerID="cri-o://9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247" gracePeriod=600 Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.411910 4829 generic.go:334] "Generic (PLEG): container finished" podID="4469e851-270f-47ee-a523-e3fb56d8201f" containerID="83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a" exitCode=0 Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.411946 4829 generic.go:334] "Generic (PLEG): container finished" podID="4469e851-270f-47ee-a523-e3fb56d8201f" containerID="b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407" exitCode=0 Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.411968 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerDied","Data":"83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a"} Oct 02 08:04:37 crc kubenswrapper[4829]: I1002 08:04:37.411998 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerDied","Data":"b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407"} Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.143129 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.249882 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4469e851-270f-47ee-a523-e3fb56d8201f-config-out\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.249965 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.250044 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.250089 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hnpl\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-kube-api-access-6hnpl\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.250126 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-thanos-prometheus-http-client-file\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.250182 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-tls-assets\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.250358 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.251021 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-config\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.251169 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.251250 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-secret-combined-ca-bundle\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.251329 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4469e851-270f-47ee-a523-e3fb56d8201f-prometheus-metric-storage-rulefiles-0\") pod \"4469e851-270f-47ee-a523-e3fb56d8201f\" (UID: \"4469e851-270f-47ee-a523-e3fb56d8201f\") " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.252935 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4469e851-270f-47ee-a523-e3fb56d8201f-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.256055 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.256668 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.259432 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-kube-api-access-6hnpl" (OuterVolumeSpecName: "kube-api-access-6hnpl") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "kube-api-access-6hnpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.259682 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.263162 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4469e851-270f-47ee-a523-e3fb56d8201f-config-out" (OuterVolumeSpecName: "config-out") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.272582 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.273388 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.275457 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-config" (OuterVolumeSpecName: "config") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.306442 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "pvc-81c3850c-e595-4f65-b168-53c52190636d". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.352985 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config" (OuterVolumeSpecName: "web-config") pod "4469e851-270f-47ee-a523-e3fb56d8201f" (UID: "4469e851-270f-47ee-a523-e3fb56d8201f"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354370 4829 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354389 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hnpl\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-kube-api-access-6hnpl\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354400 4829 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354410 4829 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4469e851-270f-47ee-a523-e3fb56d8201f-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354420 4829 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354429 4829 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-config\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354459 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") on node \"crc\" " Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354471 4829 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354482 4829 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/4469e851-270f-47ee-a523-e3fb56d8201f-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354492 4829 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4469e851-270f-47ee-a523-e3fb56d8201f-config-out\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.354501 4829 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/4469e851-270f-47ee-a523-e3fb56d8201f-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.382668 4829 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.382831 4829 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-81c3850c-e595-4f65-b168-53c52190636d" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d") on node "crc" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.423492 4829 generic.go:334] "Generic (PLEG): container finished" podID="4469e851-270f-47ee-a523-e3fb56d8201f" containerID="9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247" exitCode=0 Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.423537 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerDied","Data":"9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247"} Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.423562 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"4469e851-270f-47ee-a523-e3fb56d8201f","Type":"ContainerDied","Data":"437d10f7d98185338bb1b1d5571e80974322d31b853f19d9456c319c916194fa"} Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.423580 4829 scope.go:117] "RemoveContainer" containerID="83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.423728 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.447480 4829 scope.go:117] "RemoveContainer" containerID="9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.456621 4829 reconciler_common.go:293] "Volume detached for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") on node \"crc\" DevicePath \"\"" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.469101 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.474306 4829 scope.go:117] "RemoveContainer" containerID="b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.477935 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.492969 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.493484 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="config-reloader" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493507 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="config-reloader" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.493522 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74d0c61-084d-49d1-877a-3325875d5c15" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493533 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74d0c61-084d-49d1-877a-3325875d5c15" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.493557 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="thanos-sidecar" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493565 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="thanos-sidecar" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.493587 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="init-config-reloader" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493594 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="init-config-reloader" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.493608 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="prometheus" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493619 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="prometheus" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493837 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="config-reloader" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493864 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74d0c61-084d-49d1-877a-3325875d5c15" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493883 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="thanos-sidecar" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.493913 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" containerName="prometheus" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.495806 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.498032 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.498309 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.498313 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.498444 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.498843 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-hrzwj" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.504552 4829 scope.go:117] "RemoveContainer" containerID="50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.509308 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.512918 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.549467 4829 scope.go:117] "RemoveContainer" containerID="83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.550051 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a\": container with ID starting with 83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a not found: ID does not exist" containerID="83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550085 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a"} err="failed to get container status \"83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a\": rpc error: code = NotFound desc = could not find container \"83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a\": container with ID starting with 83b177007ac57ad08eb3e279f1534feeb8139dde955e2fdffea5976859eb5e5a not found: ID does not exist" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550109 4829 scope.go:117] "RemoveContainer" containerID="9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.550308 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247\": container with ID starting with 9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247 not found: ID does not exist" containerID="9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550324 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247"} err="failed to get container status \"9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247\": rpc error: code = NotFound desc = could not find container \"9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247\": container with ID starting with 9cf41443a2b48d82a9f600ed6f23c493f301410ca027c6191ba3d533533da247 not found: ID does not exist" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550336 4829 scope.go:117] "RemoveContainer" containerID="b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.550494 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407\": container with ID starting with b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407 not found: ID does not exist" containerID="b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550543 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407"} err="failed to get container status \"b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407\": rpc error: code = NotFound desc = could not find container \"b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407\": container with ID starting with b5949fbfdda34c2f74759f9696008c01cca43660be2b9be29f59ab59745e7407 not found: ID does not exist" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550557 4829 scope.go:117] "RemoveContainer" containerID="50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6" Oct 02 08:04:38 crc kubenswrapper[4829]: E1002 08:04:38.550699 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6\": container with ID starting with 50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6 not found: ID does not exist" containerID="50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.550733 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6"} err="failed to get container status \"50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6\": rpc error: code = NotFound desc = could not find container \"50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6\": container with ID starting with 50fa4216b673950912465998e20259a57bfb24cf45f7bd25c67a467dccaf1ff6 not found: ID does not exist" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.557925 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2de0625f-160a-4ca9-865e-35a8f782ed2d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.557969 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.558018 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.558114 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.558320 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2de0625f-160a-4ca9-865e-35a8f782ed2d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.559018 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.559071 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.559098 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2de0625f-160a-4ca9-865e-35a8f782ed2d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.559117 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.559271 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-config\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.559354 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktf5q\" (UniqueName: \"kubernetes.io/projected/2de0625f-160a-4ca9-865e-35a8f782ed2d-kube-api-access-ktf5q\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.660966 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661366 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2de0625f-160a-4ca9-865e-35a8f782ed2d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661444 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661485 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661515 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2de0625f-160a-4ca9-865e-35a8f782ed2d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661538 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661569 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-config\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661621 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktf5q\" (UniqueName: \"kubernetes.io/projected/2de0625f-160a-4ca9-865e-35a8f782ed2d-kube-api-access-ktf5q\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661653 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2de0625f-160a-4ca9-865e-35a8f782ed2d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661681 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.661734 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.663695 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/2de0625f-160a-4ca9-865e-35a8f782ed2d-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.665164 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2de0625f-160a-4ca9-865e-35a8f782ed2d-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.665573 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.668630 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.669542 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.672006 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-config\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.673091 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.673578 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2de0625f-160a-4ca9-865e-35a8f782ed2d-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.676280 4829 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.676324 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bf460093261147c62473ff98f8811184b7db6218b040c193e53546be3326fb82/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.676772 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2de0625f-160a-4ca9-865e-35a8f782ed2d-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.679172 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktf5q\" (UniqueName: \"kubernetes.io/projected/2de0625f-160a-4ca9-865e-35a8f782ed2d-kube-api-access-ktf5q\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.735043 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-81c3850c-e595-4f65-b168-53c52190636d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-81c3850c-e595-4f65-b168-53c52190636d\") pod \"prometheus-metric-storage-0\" (UID: \"2de0625f-160a-4ca9-865e-35a8f782ed2d\") " pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:38 crc kubenswrapper[4829]: I1002 08:04:38.820355 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 08:04:39 crc kubenswrapper[4829]: I1002 08:04:39.109407 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 08:04:39 crc kubenswrapper[4829]: W1002 08:04:39.110707 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2de0625f_160a_4ca9_865e_35a8f782ed2d.slice/crio-1104320dbc250b72ccd3f7f85f9cf8e2b4111aa1eaf72f897836050badbacfb7 WatchSource:0}: Error finding container 1104320dbc250b72ccd3f7f85f9cf8e2b4111aa1eaf72f897836050badbacfb7: Status 404 returned error can't find the container with id 1104320dbc250b72ccd3f7f85f9cf8e2b4111aa1eaf72f897836050badbacfb7 Oct 02 08:04:39 crc kubenswrapper[4829]: I1002 08:04:39.431950 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2de0625f-160a-4ca9-865e-35a8f782ed2d","Type":"ContainerStarted","Data":"1104320dbc250b72ccd3f7f85f9cf8e2b4111aa1eaf72f897836050badbacfb7"} Oct 02 08:04:39 crc kubenswrapper[4829]: I1002 08:04:39.478917 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4469e851-270f-47ee-a523-e3fb56d8201f" path="/var/lib/kubelet/pods/4469e851-270f-47ee-a523-e3fb56d8201f/volumes" Oct 02 08:04:43 crc kubenswrapper[4829]: I1002 08:04:43.498785 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2de0625f-160a-4ca9-865e-35a8f782ed2d","Type":"ContainerStarted","Data":"20d3b0e501721bd9742ede3dfbc04db11d15cd6a83461b3beba1316af7d2e07c"} Oct 02 08:04:54 crc kubenswrapper[4829]: E1002 08:04:54.559756 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2de0625f_160a_4ca9_865e_35a8f782ed2d.slice/crio-conmon-20d3b0e501721bd9742ede3dfbc04db11d15cd6a83461b3beba1316af7d2e07c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 08:04:54 crc kubenswrapper[4829]: I1002 08:04:54.651088 4829 generic.go:334] "Generic (PLEG): container finished" podID="2de0625f-160a-4ca9-865e-35a8f782ed2d" containerID="20d3b0e501721bd9742ede3dfbc04db11d15cd6a83461b3beba1316af7d2e07c" exitCode=0 Oct 02 08:04:54 crc kubenswrapper[4829]: I1002 08:04:54.651405 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2de0625f-160a-4ca9-865e-35a8f782ed2d","Type":"ContainerDied","Data":"20d3b0e501721bd9742ede3dfbc04db11d15cd6a83461b3beba1316af7d2e07c"} Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.329529 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.329849 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.329898 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.330608 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.330685 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" gracePeriod=600 Oct 02 08:04:55 crc kubenswrapper[4829]: E1002 08:04:55.465385 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.662371 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" exitCode=0 Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.662432 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce"} Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.662582 4829 scope.go:117] "RemoveContainer" containerID="ff7d7fa7a0802a3943548b15c46ba270f460ad42261eec82e1b15ecc455b5474" Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.663262 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:04:55 crc kubenswrapper[4829]: E1002 08:04:55.663502 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:04:55 crc kubenswrapper[4829]: I1002 08:04:55.668434 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2de0625f-160a-4ca9-865e-35a8f782ed2d","Type":"ContainerStarted","Data":"eceda64cbd2e3e8fd902bee4476ab5cb1466bc70184a28459c382c11c5998162"} Oct 02 08:04:59 crc kubenswrapper[4829]: I1002 08:04:59.747477 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2de0625f-160a-4ca9-865e-35a8f782ed2d","Type":"ContainerStarted","Data":"9124f43e352fe7e41b4d663695df5159fa322d27cb61473ac3b65a9ad9769b2c"} Oct 02 08:04:59 crc kubenswrapper[4829]: I1002 08:04:59.748924 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"2de0625f-160a-4ca9-865e-35a8f782ed2d","Type":"ContainerStarted","Data":"53d456f5316c434dadb516e1ba9c59a809305e7728063487c4730b7fa8d38c4b"} Oct 02 08:04:59 crc kubenswrapper[4829]: I1002 08:04:59.784561 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.78454186 podStartE2EDuration="21.78454186s" podCreationTimestamp="2025-10-02 08:04:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:04:59.775804298 +0000 UTC m=+2891.115452713" watchObservedRunningTime="2025-10-02 08:04:59.78454186 +0000 UTC m=+2891.124190265" Oct 02 08:05:03 crc kubenswrapper[4829]: I1002 08:05:03.821349 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 08:05:06 crc kubenswrapper[4829]: I1002 08:05:06.489745 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:05:06 crc kubenswrapper[4829]: E1002 08:05:06.490572 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:05:08 crc kubenswrapper[4829]: I1002 08:05:08.820961 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 08:05:08 crc kubenswrapper[4829]: I1002 08:05:08.831554 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 08:05:08 crc kubenswrapper[4829]: I1002 08:05:08.874946 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 08:05:19 crc kubenswrapper[4829]: I1002 08:05:19.474282 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:05:19 crc kubenswrapper[4829]: E1002 08:05:19.475550 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.128178 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.130119 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.132878 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.133032 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.133116 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-lsrdp" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.134844 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.176604 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283490 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-config-data\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283544 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283593 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283627 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283721 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283747 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283775 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283814 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ccxm\" (UniqueName: \"kubernetes.io/projected/18e0103e-1a6b-4f9c-b583-f3dc8f405137-kube-api-access-2ccxm\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.283844 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386379 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ccxm\" (UniqueName: \"kubernetes.io/projected/18e0103e-1a6b-4f9c-b583-f3dc8f405137-kube-api-access-2ccxm\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386438 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386526 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-config-data\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386559 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386606 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386640 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386733 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386759 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.386788 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.387120 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.387667 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.387867 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.388854 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.389140 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-config-data\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.396272 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.396549 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.397109 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.418854 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ccxm\" (UniqueName: \"kubernetes.io/projected/18e0103e-1a6b-4f9c-b583-f3dc8f405137-kube-api-access-2ccxm\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.432977 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.468499 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 08:05:27 crc kubenswrapper[4829]: I1002 08:05:27.854033 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 02 08:05:28 crc kubenswrapper[4829]: I1002 08:05:28.080008 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"18e0103e-1a6b-4f9c-b583-f3dc8f405137","Type":"ContainerStarted","Data":"622f7d7e84213d616a123ebc9e15c6229565096cbee3023dcf862a90c5f3a58a"} Oct 02 08:05:32 crc kubenswrapper[4829]: I1002 08:05:32.461348 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:05:32 crc kubenswrapper[4829]: E1002 08:05:32.461946 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:05:41 crc kubenswrapper[4829]: I1002 08:05:41.241930 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"18e0103e-1a6b-4f9c-b583-f3dc8f405137","Type":"ContainerStarted","Data":"7239b6eb333cd5b8373505a0b8edcd4335aa85af66728a195ad41f6dc223ba8b"} Oct 02 08:05:41 crc kubenswrapper[4829]: I1002 08:05:41.276649 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.40479536 podStartE2EDuration="15.276622276s" podCreationTimestamp="2025-10-02 08:05:26 +0000 UTC" firstStartedPulling="2025-10-02 08:05:27.857938189 +0000 UTC m=+2919.197586614" lastFinishedPulling="2025-10-02 08:05:39.729765085 +0000 UTC m=+2931.069413530" observedRunningTime="2025-10-02 08:05:41.265448474 +0000 UTC m=+2932.605096879" watchObservedRunningTime="2025-10-02 08:05:41.276622276 +0000 UTC m=+2932.616270691" Oct 02 08:05:45 crc kubenswrapper[4829]: I1002 08:05:45.461954 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:05:45 crc kubenswrapper[4829]: E1002 08:05:45.463270 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:05:59 crc kubenswrapper[4829]: I1002 08:05:59.469662 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:05:59 crc kubenswrapper[4829]: E1002 08:05:59.471479 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:06:08 crc kubenswrapper[4829]: I1002 08:06:08.794470 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-r9vxv"] Oct 02 08:06:08 crc kubenswrapper[4829]: I1002 08:06:08.798133 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:08 crc kubenswrapper[4829]: I1002 08:06:08.809532 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r9vxv"] Oct 02 08:06:08 crc kubenswrapper[4829]: I1002 08:06:08.903427 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-utilities\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:08 crc kubenswrapper[4829]: I1002 08:06:08.903524 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmm9c\" (UniqueName: \"kubernetes.io/projected/4fa88f8f-24e4-4c9b-980b-9584014bdb34-kube-api-access-gmm9c\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:08 crc kubenswrapper[4829]: I1002 08:06:08.903615 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-catalog-content\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.005491 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-utilities\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.005575 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmm9c\" (UniqueName: \"kubernetes.io/projected/4fa88f8f-24e4-4c9b-980b-9584014bdb34-kube-api-access-gmm9c\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.005647 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-catalog-content\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.006048 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-utilities\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.006135 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-catalog-content\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.030701 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmm9c\" (UniqueName: \"kubernetes.io/projected/4fa88f8f-24e4-4c9b-980b-9584014bdb34-kube-api-access-gmm9c\") pod \"community-operators-r9vxv\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.120074 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:09 crc kubenswrapper[4829]: I1002 08:06:09.693572 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-r9vxv"] Oct 02 08:06:09 crc kubenswrapper[4829]: W1002 08:06:09.704252 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fa88f8f_24e4_4c9b_980b_9584014bdb34.slice/crio-a700733d9bcb1a9c1b2239fe8c0291633c594c17d85871872bb3681ed30380f1 WatchSource:0}: Error finding container a700733d9bcb1a9c1b2239fe8c0291633c594c17d85871872bb3681ed30380f1: Status 404 returned error can't find the container with id a700733d9bcb1a9c1b2239fe8c0291633c594c17d85871872bb3681ed30380f1 Oct 02 08:06:10 crc kubenswrapper[4829]: I1002 08:06:10.589146 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerStarted","Data":"a700733d9bcb1a9c1b2239fe8c0291633c594c17d85871872bb3681ed30380f1"} Oct 02 08:06:11 crc kubenswrapper[4829]: I1002 08:06:11.603333 4829 generic.go:334] "Generic (PLEG): container finished" podID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerID="9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb" exitCode=0 Oct 02 08:06:11 crc kubenswrapper[4829]: I1002 08:06:11.603408 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerDied","Data":"9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb"} Oct 02 08:06:12 crc kubenswrapper[4829]: I1002 08:06:12.613799 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerStarted","Data":"eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d"} Oct 02 08:06:13 crc kubenswrapper[4829]: I1002 08:06:13.461014 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:06:13 crc kubenswrapper[4829]: E1002 08:06:13.461610 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:06:14 crc kubenswrapper[4829]: I1002 08:06:14.667723 4829 generic.go:334] "Generic (PLEG): container finished" podID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerID="eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d" exitCode=0 Oct 02 08:06:14 crc kubenswrapper[4829]: I1002 08:06:14.668155 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerDied","Data":"eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d"} Oct 02 08:06:15 crc kubenswrapper[4829]: I1002 08:06:15.679448 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerStarted","Data":"e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86"} Oct 02 08:06:15 crc kubenswrapper[4829]: I1002 08:06:15.705004 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-r9vxv" podStartSLOduration=4.253816032 podStartE2EDuration="7.704970914s" podCreationTimestamp="2025-10-02 08:06:08 +0000 UTC" firstStartedPulling="2025-10-02 08:06:11.606650277 +0000 UTC m=+2962.946298722" lastFinishedPulling="2025-10-02 08:06:15.057805199 +0000 UTC m=+2966.397453604" observedRunningTime="2025-10-02 08:06:15.701553136 +0000 UTC m=+2967.041201641" watchObservedRunningTime="2025-10-02 08:06:15.704970914 +0000 UTC m=+2967.044619369" Oct 02 08:06:19 crc kubenswrapper[4829]: I1002 08:06:19.120934 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:19 crc kubenswrapper[4829]: I1002 08:06:19.121761 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:19 crc kubenswrapper[4829]: I1002 08:06:19.191254 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:28 crc kubenswrapper[4829]: I1002 08:06:28.461016 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:06:28 crc kubenswrapper[4829]: E1002 08:06:28.462326 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:06:29 crc kubenswrapper[4829]: I1002 08:06:29.214306 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:29 crc kubenswrapper[4829]: I1002 08:06:29.291194 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r9vxv"] Oct 02 08:06:29 crc kubenswrapper[4829]: I1002 08:06:29.866011 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-r9vxv" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="registry-server" containerID="cri-o://e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86" gracePeriod=2 Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.408057 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.538041 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmm9c\" (UniqueName: \"kubernetes.io/projected/4fa88f8f-24e4-4c9b-980b-9584014bdb34-kube-api-access-gmm9c\") pod \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.538146 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-catalog-content\") pod \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.538354 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-utilities\") pod \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\" (UID: \"4fa88f8f-24e4-4c9b-980b-9584014bdb34\") " Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.539701 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-utilities" (OuterVolumeSpecName: "utilities") pod "4fa88f8f-24e4-4c9b-980b-9584014bdb34" (UID: "4fa88f8f-24e4-4c9b-980b-9584014bdb34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.550294 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa88f8f-24e4-4c9b-980b-9584014bdb34-kube-api-access-gmm9c" (OuterVolumeSpecName: "kube-api-access-gmm9c") pod "4fa88f8f-24e4-4c9b-980b-9584014bdb34" (UID: "4fa88f8f-24e4-4c9b-980b-9584014bdb34"). InnerVolumeSpecName "kube-api-access-gmm9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.609070 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fa88f8f-24e4-4c9b-980b-9584014bdb34" (UID: "4fa88f8f-24e4-4c9b-980b-9584014bdb34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.640553 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.640835 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmm9c\" (UniqueName: \"kubernetes.io/projected/4fa88f8f-24e4-4c9b-980b-9584014bdb34-kube-api-access-gmm9c\") on node \"crc\" DevicePath \"\"" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.640850 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fa88f8f-24e4-4c9b-980b-9584014bdb34-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.902875 4829 generic.go:334] "Generic (PLEG): container finished" podID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerID="e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86" exitCode=0 Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.902961 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerDied","Data":"e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86"} Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.903021 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-r9vxv" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.903288 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-r9vxv" event={"ID":"4fa88f8f-24e4-4c9b-980b-9584014bdb34","Type":"ContainerDied","Data":"a700733d9bcb1a9c1b2239fe8c0291633c594c17d85871872bb3681ed30380f1"} Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.903367 4829 scope.go:117] "RemoveContainer" containerID="e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.964814 4829 scope.go:117] "RemoveContainer" containerID="eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d" Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.967583 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-r9vxv"] Oct 02 08:06:30 crc kubenswrapper[4829]: I1002 08:06:30.987591 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-r9vxv"] Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.017106 4829 scope.go:117] "RemoveContainer" containerID="9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.073822 4829 scope.go:117] "RemoveContainer" containerID="e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86" Oct 02 08:06:31 crc kubenswrapper[4829]: E1002 08:06:31.074519 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86\": container with ID starting with e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86 not found: ID does not exist" containerID="e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.074585 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86"} err="failed to get container status \"e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86\": rpc error: code = NotFound desc = could not find container \"e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86\": container with ID starting with e6e4dfd1ef5cd3d842b6ca6e1abaacc4472d97664370f8b273016943acceca86 not found: ID does not exist" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.074629 4829 scope.go:117] "RemoveContainer" containerID="eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d" Oct 02 08:06:31 crc kubenswrapper[4829]: E1002 08:06:31.075173 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d\": container with ID starting with eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d not found: ID does not exist" containerID="eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.075281 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d"} err="failed to get container status \"eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d\": rpc error: code = NotFound desc = could not find container \"eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d\": container with ID starting with eeaaba283efa62b6ddcd8b77248b384386250f9966767ef7357631cdd5e25a0d not found: ID does not exist" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.075343 4829 scope.go:117] "RemoveContainer" containerID="9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb" Oct 02 08:06:31 crc kubenswrapper[4829]: E1002 08:06:31.075828 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb\": container with ID starting with 9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb not found: ID does not exist" containerID="9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.075878 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb"} err="failed to get container status \"9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb\": rpc error: code = NotFound desc = could not find container \"9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb\": container with ID starting with 9e96712402cc5c54d13fc2be609fcb4651f3f028e4c90d14425c6ca746bbc7cb not found: ID does not exist" Oct 02 08:06:31 crc kubenswrapper[4829]: I1002 08:06:31.476617 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" path="/var/lib/kubelet/pods/4fa88f8f-24e4-4c9b-980b-9584014bdb34/volumes" Oct 02 08:06:42 crc kubenswrapper[4829]: I1002 08:06:42.462366 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:06:42 crc kubenswrapper[4829]: E1002 08:06:42.463732 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:06:56 crc kubenswrapper[4829]: I1002 08:06:56.461891 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:06:56 crc kubenswrapper[4829]: E1002 08:06:56.462907 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:07:08 crc kubenswrapper[4829]: I1002 08:07:08.461182 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:07:08 crc kubenswrapper[4829]: E1002 08:07:08.462107 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:07:22 crc kubenswrapper[4829]: I1002 08:07:22.461675 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:07:22 crc kubenswrapper[4829]: E1002 08:07:22.462814 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:07:37 crc kubenswrapper[4829]: I1002 08:07:37.461747 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:07:37 crc kubenswrapper[4829]: E1002 08:07:37.463027 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:07:51 crc kubenswrapper[4829]: I1002 08:07:51.461142 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:07:51 crc kubenswrapper[4829]: E1002 08:07:51.461996 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:08:06 crc kubenswrapper[4829]: I1002 08:08:06.461495 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:08:06 crc kubenswrapper[4829]: E1002 08:08:06.464195 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:08:17 crc kubenswrapper[4829]: I1002 08:08:17.461986 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:08:17 crc kubenswrapper[4829]: E1002 08:08:17.463297 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:08:29 crc kubenswrapper[4829]: I1002 08:08:29.475713 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:08:29 crc kubenswrapper[4829]: E1002 08:08:29.477051 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.693682 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bs8bh"] Oct 02 08:08:38 crc kubenswrapper[4829]: E1002 08:08:38.695023 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="extract-utilities" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.695047 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="extract-utilities" Oct 02 08:08:38 crc kubenswrapper[4829]: E1002 08:08:38.695079 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="extract-content" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.695092 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="extract-content" Oct 02 08:08:38 crc kubenswrapper[4829]: E1002 08:08:38.695165 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="registry-server" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.695179 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="registry-server" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.695536 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa88f8f-24e4-4c9b-980b-9584014bdb34" containerName="registry-server" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.698192 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.708063 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bs8bh"] Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.767952 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-utilities\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.768441 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-catalog-content\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.768618 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8zhn\" (UniqueName: \"kubernetes.io/projected/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-kube-api-access-c8zhn\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.870006 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-catalog-content\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.870101 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8zhn\" (UniqueName: \"kubernetes.io/projected/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-kube-api-access-c8zhn\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.870162 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-utilities\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.870505 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-catalog-content\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.871142 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-utilities\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:38 crc kubenswrapper[4829]: I1002 08:08:38.894651 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8zhn\" (UniqueName: \"kubernetes.io/projected/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-kube-api-access-c8zhn\") pod \"redhat-marketplace-bs8bh\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:39 crc kubenswrapper[4829]: I1002 08:08:39.040620 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:39 crc kubenswrapper[4829]: I1002 08:08:39.558216 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bs8bh"] Oct 02 08:08:40 crc kubenswrapper[4829]: I1002 08:08:40.540700 4829 generic.go:334] "Generic (PLEG): container finished" podID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerID="7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76" exitCode=0 Oct 02 08:08:40 crc kubenswrapper[4829]: I1002 08:08:40.540780 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bs8bh" event={"ID":"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4","Type":"ContainerDied","Data":"7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76"} Oct 02 08:08:40 crc kubenswrapper[4829]: I1002 08:08:40.541004 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bs8bh" event={"ID":"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4","Type":"ContainerStarted","Data":"e833923e96dc52dba3bd34e130bed2ed1742784de72a8b5fc0066c3a2e839845"} Oct 02 08:08:40 crc kubenswrapper[4829]: I1002 08:08:40.544785 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:08:42 crc kubenswrapper[4829]: I1002 08:08:42.461292 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:08:42 crc kubenswrapper[4829]: E1002 08:08:42.462525 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:08:42 crc kubenswrapper[4829]: I1002 08:08:42.568289 4829 generic.go:334] "Generic (PLEG): container finished" podID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerID="a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d" exitCode=0 Oct 02 08:08:42 crc kubenswrapper[4829]: I1002 08:08:42.568364 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bs8bh" event={"ID":"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4","Type":"ContainerDied","Data":"a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d"} Oct 02 08:08:43 crc kubenswrapper[4829]: I1002 08:08:43.583954 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bs8bh" event={"ID":"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4","Type":"ContainerStarted","Data":"a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3"} Oct 02 08:08:43 crc kubenswrapper[4829]: I1002 08:08:43.614476 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bs8bh" podStartSLOduration=3.055798368 podStartE2EDuration="5.614457665s" podCreationTimestamp="2025-10-02 08:08:38 +0000 UTC" firstStartedPulling="2025-10-02 08:08:40.54441295 +0000 UTC m=+3111.884061375" lastFinishedPulling="2025-10-02 08:08:43.103072257 +0000 UTC m=+3114.442720672" observedRunningTime="2025-10-02 08:08:43.610068667 +0000 UTC m=+3114.949717112" watchObservedRunningTime="2025-10-02 08:08:43.614457665 +0000 UTC m=+3114.954106070" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.644123 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kc7w9"] Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.646508 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.661213 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kc7w9"] Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.804058 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-catalog-content\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.805329 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkzzd\" (UniqueName: \"kubernetes.io/projected/e8da322f-391c-494f-ad81-239cc6a1f8db-kube-api-access-qkzzd\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.805392 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-utilities\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.907915 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-catalog-content\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.908108 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-utilities\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.908142 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkzzd\" (UniqueName: \"kubernetes.io/projected/e8da322f-391c-494f-ad81-239cc6a1f8db-kube-api-access-qkzzd\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.908735 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-utilities\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.910562 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-catalog-content\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.933702 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkzzd\" (UniqueName: \"kubernetes.io/projected/e8da322f-391c-494f-ad81-239cc6a1f8db-kube-api-access-qkzzd\") pod \"certified-operators-kc7w9\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:48 crc kubenswrapper[4829]: I1002 08:08:48.986011 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:49 crc kubenswrapper[4829]: I1002 08:08:49.040800 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:49 crc kubenswrapper[4829]: I1002 08:08:49.040959 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:49 crc kubenswrapper[4829]: I1002 08:08:49.124987 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:49 crc kubenswrapper[4829]: I1002 08:08:49.496802 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kc7w9"] Oct 02 08:08:49 crc kubenswrapper[4829]: I1002 08:08:49.663722 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerStarted","Data":"25850642987b8f73559eef79e28a6d64eb0087e917f75062d40396b087fbb2fb"} Oct 02 08:08:49 crc kubenswrapper[4829]: I1002 08:08:49.752047 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:50 crc kubenswrapper[4829]: I1002 08:08:50.675583 4829 generic.go:334] "Generic (PLEG): container finished" podID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerID="2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168" exitCode=0 Oct 02 08:08:50 crc kubenswrapper[4829]: I1002 08:08:50.675691 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerDied","Data":"2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168"} Oct 02 08:08:51 crc kubenswrapper[4829]: I1002 08:08:51.407950 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bs8bh"] Oct 02 08:08:51 crc kubenswrapper[4829]: I1002 08:08:51.687938 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerStarted","Data":"074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5"} Oct 02 08:08:52 crc kubenswrapper[4829]: I1002 08:08:52.697281 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bs8bh" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="registry-server" containerID="cri-o://a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3" gracePeriod=2 Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.326461 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.507600 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-catalog-content\") pod \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.507749 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8zhn\" (UniqueName: \"kubernetes.io/projected/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-kube-api-access-c8zhn\") pod \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.507821 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-utilities\") pod \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\" (UID: \"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4\") " Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.508559 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-utilities" (OuterVolumeSpecName: "utilities") pod "3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" (UID: "3cbc73e9-14ee-4e72-8812-8fcaee24cdb4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.519605 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-kube-api-access-c8zhn" (OuterVolumeSpecName: "kube-api-access-c8zhn") pod "3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" (UID: "3cbc73e9-14ee-4e72-8812-8fcaee24cdb4"). InnerVolumeSpecName "kube-api-access-c8zhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.521945 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" (UID: "3cbc73e9-14ee-4e72-8812-8fcaee24cdb4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.610006 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8zhn\" (UniqueName: \"kubernetes.io/projected/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-kube-api-access-c8zhn\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.610040 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.610054 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.712709 4829 generic.go:334] "Generic (PLEG): container finished" podID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerID="a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3" exitCode=0 Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.712795 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bs8bh" event={"ID":"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4","Type":"ContainerDied","Data":"a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3"} Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.712826 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bs8bh" event={"ID":"3cbc73e9-14ee-4e72-8812-8fcaee24cdb4","Type":"ContainerDied","Data":"e833923e96dc52dba3bd34e130bed2ed1742784de72a8b5fc0066c3a2e839845"} Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.712846 4829 scope.go:117] "RemoveContainer" containerID="a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.712862 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bs8bh" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.733627 4829 generic.go:334] "Generic (PLEG): container finished" podID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerID="074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5" exitCode=0 Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.734055 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerDied","Data":"074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5"} Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.789153 4829 scope.go:117] "RemoveContainer" containerID="a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.814476 4829 scope.go:117] "RemoveContainer" containerID="7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.814640 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bs8bh"] Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.828530 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bs8bh"] Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.892487 4829 scope.go:117] "RemoveContainer" containerID="a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3" Oct 02 08:08:53 crc kubenswrapper[4829]: E1002 08:08:53.893037 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3\": container with ID starting with a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3 not found: ID does not exist" containerID="a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.893075 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3"} err="failed to get container status \"a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3\": rpc error: code = NotFound desc = could not find container \"a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3\": container with ID starting with a16c665c9920cd994f262ea6caab5d7018194c5581eb0377a86e7595cbc552d3 not found: ID does not exist" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.893108 4829 scope.go:117] "RemoveContainer" containerID="a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d" Oct 02 08:08:53 crc kubenswrapper[4829]: E1002 08:08:53.893516 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d\": container with ID starting with a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d not found: ID does not exist" containerID="a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.893588 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d"} err="failed to get container status \"a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d\": rpc error: code = NotFound desc = could not find container \"a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d\": container with ID starting with a55d76cf57e2f7db7f3498c9f30b5bd22faec842fd4a82967e7de32b9fbd825d not found: ID does not exist" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.893641 4829 scope.go:117] "RemoveContainer" containerID="7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76" Oct 02 08:08:53 crc kubenswrapper[4829]: E1002 08:08:53.894043 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76\": container with ID starting with 7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76 not found: ID does not exist" containerID="7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76" Oct 02 08:08:53 crc kubenswrapper[4829]: I1002 08:08:53.894072 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76"} err="failed to get container status \"7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76\": rpc error: code = NotFound desc = could not find container \"7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76\": container with ID starting with 7c310df0a37feb216a8c890f993fe7ef4f4c7800092a206105e5f9ab3cafab76 not found: ID does not exist" Oct 02 08:08:55 crc kubenswrapper[4829]: I1002 08:08:55.479521 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" path="/var/lib/kubelet/pods/3cbc73e9-14ee-4e72-8812-8fcaee24cdb4/volumes" Oct 02 08:08:55 crc kubenswrapper[4829]: I1002 08:08:55.763714 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerStarted","Data":"e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3"} Oct 02 08:08:55 crc kubenswrapper[4829]: I1002 08:08:55.788793 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kc7w9" podStartSLOduration=3.978889155 podStartE2EDuration="7.788773031s" podCreationTimestamp="2025-10-02 08:08:48 +0000 UTC" firstStartedPulling="2025-10-02 08:08:50.677492998 +0000 UTC m=+3122.017141403" lastFinishedPulling="2025-10-02 08:08:54.487376844 +0000 UTC m=+3125.827025279" observedRunningTime="2025-10-02 08:08:55.77953332 +0000 UTC m=+3127.119181765" watchObservedRunningTime="2025-10-02 08:08:55.788773031 +0000 UTC m=+3127.128421446" Oct 02 08:08:56 crc kubenswrapper[4829]: I1002 08:08:56.461451 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:08:56 crc kubenswrapper[4829]: E1002 08:08:56.461786 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:08:58 crc kubenswrapper[4829]: I1002 08:08:58.986768 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:58 crc kubenswrapper[4829]: I1002 08:08:58.987403 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:59 crc kubenswrapper[4829]: I1002 08:08:59.063345 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:08:59 crc kubenswrapper[4829]: I1002 08:08:59.869679 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:09:00 crc kubenswrapper[4829]: I1002 08:09:00.215687 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kc7w9"] Oct 02 08:09:01 crc kubenswrapper[4829]: I1002 08:09:01.829828 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kc7w9" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="registry-server" containerID="cri-o://e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3" gracePeriod=2 Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.326083 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.401535 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-catalog-content\") pod \"e8da322f-391c-494f-ad81-239cc6a1f8db\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.401657 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkzzd\" (UniqueName: \"kubernetes.io/projected/e8da322f-391c-494f-ad81-239cc6a1f8db-kube-api-access-qkzzd\") pod \"e8da322f-391c-494f-ad81-239cc6a1f8db\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.401720 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-utilities\") pod \"e8da322f-391c-494f-ad81-239cc6a1f8db\" (UID: \"e8da322f-391c-494f-ad81-239cc6a1f8db\") " Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.403092 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-utilities" (OuterVolumeSpecName: "utilities") pod "e8da322f-391c-494f-ad81-239cc6a1f8db" (UID: "e8da322f-391c-494f-ad81-239cc6a1f8db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.410529 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8da322f-391c-494f-ad81-239cc6a1f8db-kube-api-access-qkzzd" (OuterVolumeSpecName: "kube-api-access-qkzzd") pod "e8da322f-391c-494f-ad81-239cc6a1f8db" (UID: "e8da322f-391c-494f-ad81-239cc6a1f8db"). InnerVolumeSpecName "kube-api-access-qkzzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.462047 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8da322f-391c-494f-ad81-239cc6a1f8db" (UID: "e8da322f-391c-494f-ad81-239cc6a1f8db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.505411 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.505458 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkzzd\" (UniqueName: \"kubernetes.io/projected/e8da322f-391c-494f-ad81-239cc6a1f8db-kube-api-access-qkzzd\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.505479 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8da322f-391c-494f-ad81-239cc6a1f8db-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.844453 4829 generic.go:334] "Generic (PLEG): container finished" podID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerID="e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3" exitCode=0 Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.844543 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc7w9" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.844539 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerDied","Data":"e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3"} Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.846361 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc7w9" event={"ID":"e8da322f-391c-494f-ad81-239cc6a1f8db","Type":"ContainerDied","Data":"25850642987b8f73559eef79e28a6d64eb0087e917f75062d40396b087fbb2fb"} Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.846408 4829 scope.go:117] "RemoveContainer" containerID="e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.887082 4829 scope.go:117] "RemoveContainer" containerID="074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.909169 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kc7w9"] Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.922078 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kc7w9"] Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.934262 4829 scope.go:117] "RemoveContainer" containerID="2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.984643 4829 scope.go:117] "RemoveContainer" containerID="e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3" Oct 02 08:09:02 crc kubenswrapper[4829]: E1002 08:09:02.988522 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3\": container with ID starting with e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3 not found: ID does not exist" containerID="e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.988600 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3"} err="failed to get container status \"e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3\": rpc error: code = NotFound desc = could not find container \"e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3\": container with ID starting with e301f513b214e9e90d3f308d4416eec0c3ca1c7c156e1ed19da68be8c321beb3 not found: ID does not exist" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.988635 4829 scope.go:117] "RemoveContainer" containerID="074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5" Oct 02 08:09:02 crc kubenswrapper[4829]: E1002 08:09:02.993846 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5\": container with ID starting with 074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5 not found: ID does not exist" containerID="074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.993897 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5"} err="failed to get container status \"074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5\": rpc error: code = NotFound desc = could not find container \"074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5\": container with ID starting with 074f1dc62c9ae315277c6a574914293cfa5a34f615690d66fa3250f242bd23c5 not found: ID does not exist" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.993930 4829 scope.go:117] "RemoveContainer" containerID="2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168" Oct 02 08:09:02 crc kubenswrapper[4829]: E1002 08:09:02.994444 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168\": container with ID starting with 2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168 not found: ID does not exist" containerID="2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168" Oct 02 08:09:02 crc kubenswrapper[4829]: I1002 08:09:02.994468 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168"} err="failed to get container status \"2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168\": rpc error: code = NotFound desc = could not find container \"2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168\": container with ID starting with 2706553420e4f6260dfc18dbf2d4a84711a704dec911fb0258b0794a63bc5168 not found: ID does not exist" Oct 02 08:09:03 crc kubenswrapper[4829]: I1002 08:09:03.480700 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" path="/var/lib/kubelet/pods/e8da322f-391c-494f-ad81-239cc6a1f8db/volumes" Oct 02 08:09:07 crc kubenswrapper[4829]: I1002 08:09:07.461869 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:09:07 crc kubenswrapper[4829]: E1002 08:09:07.463100 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:09:19 crc kubenswrapper[4829]: I1002 08:09:19.473025 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:09:19 crc kubenswrapper[4829]: E1002 08:09:19.474123 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:09:31 crc kubenswrapper[4829]: I1002 08:09:31.460723 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:09:31 crc kubenswrapper[4829]: E1002 08:09:31.461518 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.800484 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5zqh6"] Oct 02 08:09:40 crc kubenswrapper[4829]: E1002 08:09:40.801523 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="registry-server" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801541 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="registry-server" Oct 02 08:09:40 crc kubenswrapper[4829]: E1002 08:09:40.801573 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="registry-server" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801581 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="registry-server" Oct 02 08:09:40 crc kubenswrapper[4829]: E1002 08:09:40.801604 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="extract-utilities" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801613 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="extract-utilities" Oct 02 08:09:40 crc kubenswrapper[4829]: E1002 08:09:40.801639 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="extract-content" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801647 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="extract-content" Oct 02 08:09:40 crc kubenswrapper[4829]: E1002 08:09:40.801665 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="extract-utilities" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801674 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="extract-utilities" Oct 02 08:09:40 crc kubenswrapper[4829]: E1002 08:09:40.801692 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="extract-content" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801701 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="extract-content" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801942 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cbc73e9-14ee-4e72-8812-8fcaee24cdb4" containerName="registry-server" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.801966 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8da322f-391c-494f-ad81-239cc6a1f8db" containerName="registry-server" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.807970 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.826261 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zqh6"] Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.974533 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-catalog-content\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.974638 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-utilities\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:40 crc kubenswrapper[4829]: I1002 08:09:40.974701 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ngf\" (UniqueName: \"kubernetes.io/projected/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-kube-api-access-s4ngf\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.076498 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-utilities\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.076644 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ngf\" (UniqueName: \"kubernetes.io/projected/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-kube-api-access-s4ngf\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.076943 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-utilities\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.077191 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-catalog-content\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.077561 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-catalog-content\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.098126 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ngf\" (UniqueName: \"kubernetes.io/projected/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-kube-api-access-s4ngf\") pod \"redhat-operators-5zqh6\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.133485 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:41 crc kubenswrapper[4829]: I1002 08:09:41.616971 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5zqh6"] Oct 02 08:09:42 crc kubenswrapper[4829]: I1002 08:09:42.337981 4829 generic.go:334] "Generic (PLEG): container finished" podID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerID="e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab" exitCode=0 Oct 02 08:09:42 crc kubenswrapper[4829]: I1002 08:09:42.338100 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerDied","Data":"e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab"} Oct 02 08:09:42 crc kubenswrapper[4829]: I1002 08:09:42.338269 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerStarted","Data":"5a167ceb33a2533682f23ca06f235fe29912df5331bd0a60707273412cd4a9e0"} Oct 02 08:09:43 crc kubenswrapper[4829]: I1002 08:09:43.348730 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerStarted","Data":"f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88"} Oct 02 08:09:46 crc kubenswrapper[4829]: I1002 08:09:46.461890 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:09:46 crc kubenswrapper[4829]: E1002 08:09:46.463143 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:09:47 crc kubenswrapper[4829]: I1002 08:09:47.405737 4829 generic.go:334] "Generic (PLEG): container finished" podID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerID="f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88" exitCode=0 Oct 02 08:09:47 crc kubenswrapper[4829]: I1002 08:09:47.405787 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerDied","Data":"f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88"} Oct 02 08:09:48 crc kubenswrapper[4829]: I1002 08:09:48.423567 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerStarted","Data":"a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c"} Oct 02 08:09:48 crc kubenswrapper[4829]: I1002 08:09:48.463530 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5zqh6" podStartSLOduration=2.96924796 podStartE2EDuration="8.4634991s" podCreationTimestamp="2025-10-02 08:09:40 +0000 UTC" firstStartedPulling="2025-10-02 08:09:42.340446585 +0000 UTC m=+3173.680094990" lastFinishedPulling="2025-10-02 08:09:47.834697725 +0000 UTC m=+3179.174346130" observedRunningTime="2025-10-02 08:09:48.449990216 +0000 UTC m=+3179.789638661" watchObservedRunningTime="2025-10-02 08:09:48.4634991 +0000 UTC m=+3179.803147545" Oct 02 08:09:51 crc kubenswrapper[4829]: I1002 08:09:51.134630 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:51 crc kubenswrapper[4829]: I1002 08:09:51.135457 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:09:52 crc kubenswrapper[4829]: I1002 08:09:52.213824 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5zqh6" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="registry-server" probeResult="failure" output=< Oct 02 08:09:52 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:09:52 crc kubenswrapper[4829]: > Oct 02 08:09:56 crc kubenswrapper[4829]: E1002 08:09:56.240582 4829 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:47644->38.102.83.200:43623: write tcp 38.102.83.200:47644->38.102.83.200:43623: write: broken pipe Oct 02 08:09:59 crc kubenswrapper[4829]: I1002 08:09:59.474444 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:10:00 crc kubenswrapper[4829]: I1002 08:10:00.578002 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"628e791b8bad71add3acfd79a863f9f6417cad671ec5f9feb48a872b7a63ccaa"} Oct 02 08:10:01 crc kubenswrapper[4829]: I1002 08:10:01.195350 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:10:01 crc kubenswrapper[4829]: I1002 08:10:01.255823 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:10:01 crc kubenswrapper[4829]: I1002 08:10:01.441023 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zqh6"] Oct 02 08:10:02 crc kubenswrapper[4829]: I1002 08:10:02.610080 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5zqh6" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="registry-server" containerID="cri-o://a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c" gracePeriod=2 Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.177734 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.271998 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4ngf\" (UniqueName: \"kubernetes.io/projected/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-kube-api-access-s4ngf\") pod \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.272541 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-catalog-content\") pod \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.272725 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-utilities\") pod \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\" (UID: \"b7437bfb-7fa1-4e7d-9cb5-173b40f65847\") " Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.274085 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-utilities" (OuterVolumeSpecName: "utilities") pod "b7437bfb-7fa1-4e7d-9cb5-173b40f65847" (UID: "b7437bfb-7fa1-4e7d-9cb5-173b40f65847"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.281664 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-kube-api-access-s4ngf" (OuterVolumeSpecName: "kube-api-access-s4ngf") pod "b7437bfb-7fa1-4e7d-9cb5-173b40f65847" (UID: "b7437bfb-7fa1-4e7d-9cb5-173b40f65847"). InnerVolumeSpecName "kube-api-access-s4ngf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.366012 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7437bfb-7fa1-4e7d-9cb5-173b40f65847" (UID: "b7437bfb-7fa1-4e7d-9cb5-173b40f65847"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.377908 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4ngf\" (UniqueName: \"kubernetes.io/projected/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-kube-api-access-s4ngf\") on node \"crc\" DevicePath \"\"" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.377966 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.377993 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7437bfb-7fa1-4e7d-9cb5-173b40f65847-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.625577 4829 generic.go:334] "Generic (PLEG): container finished" podID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerID="a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c" exitCode=0 Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.625620 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerDied","Data":"a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c"} Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.625654 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5zqh6" event={"ID":"b7437bfb-7fa1-4e7d-9cb5-173b40f65847","Type":"ContainerDied","Data":"5a167ceb33a2533682f23ca06f235fe29912df5331bd0a60707273412cd4a9e0"} Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.625674 4829 scope.go:117] "RemoveContainer" containerID="a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.625769 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5zqh6" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.654110 4829 scope.go:117] "RemoveContainer" containerID="f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.659452 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5zqh6"] Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.673273 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5zqh6"] Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.679417 4829 scope.go:117] "RemoveContainer" containerID="e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.747538 4829 scope.go:117] "RemoveContainer" containerID="a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c" Oct 02 08:10:03 crc kubenswrapper[4829]: E1002 08:10:03.748501 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c\": container with ID starting with a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c not found: ID does not exist" containerID="a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.748548 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c"} err="failed to get container status \"a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c\": rpc error: code = NotFound desc = could not find container \"a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c\": container with ID starting with a092529e84bf05e43af2b82e7fe539aa604556cf0e7a7885acd538d1bf21a54c not found: ID does not exist" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.748575 4829 scope.go:117] "RemoveContainer" containerID="f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88" Oct 02 08:10:03 crc kubenswrapper[4829]: E1002 08:10:03.751000 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88\": container with ID starting with f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88 not found: ID does not exist" containerID="f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.751076 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88"} err="failed to get container status \"f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88\": rpc error: code = NotFound desc = could not find container \"f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88\": container with ID starting with f4ad94056f474bee001d86a1776e3fc58cd0876018f473b2f4e68052c04d5d88 not found: ID does not exist" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.751137 4829 scope.go:117] "RemoveContainer" containerID="e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab" Oct 02 08:10:03 crc kubenswrapper[4829]: E1002 08:10:03.751624 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab\": container with ID starting with e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab not found: ID does not exist" containerID="e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab" Oct 02 08:10:03 crc kubenswrapper[4829]: I1002 08:10:03.751655 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab"} err="failed to get container status \"e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab\": rpc error: code = NotFound desc = could not find container \"e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab\": container with ID starting with e3c3006329c906edf5263372e27f4599da797821da12bf9361994bb1accf40ab not found: ID does not exist" Oct 02 08:10:05 crc kubenswrapper[4829]: I1002 08:10:05.473910 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" path="/var/lib/kubelet/pods/b7437bfb-7fa1-4e7d-9cb5-173b40f65847/volumes" Oct 02 08:12:25 crc kubenswrapper[4829]: I1002 08:12:25.329096 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:12:25 crc kubenswrapper[4829]: I1002 08:12:25.329726 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:12:55 crc kubenswrapper[4829]: I1002 08:12:55.329840 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:12:55 crc kubenswrapper[4829]: I1002 08:12:55.330532 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:13:25 crc kubenswrapper[4829]: I1002 08:13:25.329283 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:13:25 crc kubenswrapper[4829]: I1002 08:13:25.329786 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:13:25 crc kubenswrapper[4829]: I1002 08:13:25.329826 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:13:25 crc kubenswrapper[4829]: I1002 08:13:25.330349 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"628e791b8bad71add3acfd79a863f9f6417cad671ec5f9feb48a872b7a63ccaa"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:13:25 crc kubenswrapper[4829]: I1002 08:13:25.330400 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://628e791b8bad71add3acfd79a863f9f6417cad671ec5f9feb48a872b7a63ccaa" gracePeriod=600 Oct 02 08:13:26 crc kubenswrapper[4829]: I1002 08:13:26.109190 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="628e791b8bad71add3acfd79a863f9f6417cad671ec5f9feb48a872b7a63ccaa" exitCode=0 Oct 02 08:13:26 crc kubenswrapper[4829]: I1002 08:13:26.109279 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"628e791b8bad71add3acfd79a863f9f6417cad671ec5f9feb48a872b7a63ccaa"} Oct 02 08:13:26 crc kubenswrapper[4829]: I1002 08:13:26.109572 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3"} Oct 02 08:13:26 crc kubenswrapper[4829]: I1002 08:13:26.109598 4829 scope.go:117] "RemoveContainer" containerID="e05b2dcb50c2719a20d5a704ce4a2760713b1df024a6d99ab358fda7aa2402ce" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.206776 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj"] Oct 02 08:15:00 crc kubenswrapper[4829]: E1002 08:15:00.209511 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="extract-utilities" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.209672 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="extract-utilities" Oct 02 08:15:00 crc kubenswrapper[4829]: E1002 08:15:00.209853 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="registry-server" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.209964 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="registry-server" Oct 02 08:15:00 crc kubenswrapper[4829]: E1002 08:15:00.210088 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="extract-content" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.210203 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="extract-content" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.210716 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7437bfb-7fa1-4e7d-9cb5-173b40f65847" containerName="registry-server" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.212036 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.214918 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.215854 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.218970 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj"] Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.373488 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d17916ae-ef45-4999-8b6d-42f8476721e9-config-volume\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.373605 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz6r4\" (UniqueName: \"kubernetes.io/projected/d17916ae-ef45-4999-8b6d-42f8476721e9-kube-api-access-vz6r4\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.373686 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d17916ae-ef45-4999-8b6d-42f8476721e9-secret-volume\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.477954 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d17916ae-ef45-4999-8b6d-42f8476721e9-config-volume\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.478063 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz6r4\" (UniqueName: \"kubernetes.io/projected/d17916ae-ef45-4999-8b6d-42f8476721e9-kube-api-access-vz6r4\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.478140 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d17916ae-ef45-4999-8b6d-42f8476721e9-secret-volume\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.479384 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d17916ae-ef45-4999-8b6d-42f8476721e9-config-volume\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.484887 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d17916ae-ef45-4999-8b6d-42f8476721e9-secret-volume\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.494728 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz6r4\" (UniqueName: \"kubernetes.io/projected/d17916ae-ef45-4999-8b6d-42f8476721e9-kube-api-access-vz6r4\") pod \"collect-profiles-29323215-qh6nj\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:00 crc kubenswrapper[4829]: I1002 08:15:00.552769 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:01 crc kubenswrapper[4829]: I1002 08:15:01.035021 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj"] Oct 02 08:15:01 crc kubenswrapper[4829]: I1002 08:15:01.265141 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" event={"ID":"d17916ae-ef45-4999-8b6d-42f8476721e9","Type":"ContainerStarted","Data":"ff7ebdfe8cdc5a0d46679c516037aab1aeb65178bf8de4a0976ab5e1c8e6b692"} Oct 02 08:15:01 crc kubenswrapper[4829]: I1002 08:15:01.265200 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" event={"ID":"d17916ae-ef45-4999-8b6d-42f8476721e9","Type":"ContainerStarted","Data":"401af2ff670333c550d972fd805237686195ed1354e47bd727544f98def4b907"} Oct 02 08:15:01 crc kubenswrapper[4829]: I1002 08:15:01.282638 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" podStartSLOduration=1.2826187820000001 podStartE2EDuration="1.282618782s" podCreationTimestamp="2025-10-02 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:15:01.277332116 +0000 UTC m=+3492.616980511" watchObservedRunningTime="2025-10-02 08:15:01.282618782 +0000 UTC m=+3492.622267187" Oct 02 08:15:02 crc kubenswrapper[4829]: I1002 08:15:02.278871 4829 generic.go:334] "Generic (PLEG): container finished" podID="d17916ae-ef45-4999-8b6d-42f8476721e9" containerID="ff7ebdfe8cdc5a0d46679c516037aab1aeb65178bf8de4a0976ab5e1c8e6b692" exitCode=0 Oct 02 08:15:02 crc kubenswrapper[4829]: I1002 08:15:02.278918 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" event={"ID":"d17916ae-ef45-4999-8b6d-42f8476721e9","Type":"ContainerDied","Data":"ff7ebdfe8cdc5a0d46679c516037aab1aeb65178bf8de4a0976ab5e1c8e6b692"} Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.787990 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.951018 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d17916ae-ef45-4999-8b6d-42f8476721e9-config-volume\") pod \"d17916ae-ef45-4999-8b6d-42f8476721e9\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.951078 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz6r4\" (UniqueName: \"kubernetes.io/projected/d17916ae-ef45-4999-8b6d-42f8476721e9-kube-api-access-vz6r4\") pod \"d17916ae-ef45-4999-8b6d-42f8476721e9\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.951116 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d17916ae-ef45-4999-8b6d-42f8476721e9-secret-volume\") pod \"d17916ae-ef45-4999-8b6d-42f8476721e9\" (UID: \"d17916ae-ef45-4999-8b6d-42f8476721e9\") " Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.952455 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d17916ae-ef45-4999-8b6d-42f8476721e9-config-volume" (OuterVolumeSpecName: "config-volume") pod "d17916ae-ef45-4999-8b6d-42f8476721e9" (UID: "d17916ae-ef45-4999-8b6d-42f8476721e9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.960019 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17916ae-ef45-4999-8b6d-42f8476721e9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d17916ae-ef45-4999-8b6d-42f8476721e9" (UID: "d17916ae-ef45-4999-8b6d-42f8476721e9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:15:03 crc kubenswrapper[4829]: I1002 08:15:03.960893 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17916ae-ef45-4999-8b6d-42f8476721e9-kube-api-access-vz6r4" (OuterVolumeSpecName: "kube-api-access-vz6r4") pod "d17916ae-ef45-4999-8b6d-42f8476721e9" (UID: "d17916ae-ef45-4999-8b6d-42f8476721e9"). InnerVolumeSpecName "kube-api-access-vz6r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.054337 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d17916ae-ef45-4999-8b6d-42f8476721e9-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.054369 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz6r4\" (UniqueName: \"kubernetes.io/projected/d17916ae-ef45-4999-8b6d-42f8476721e9-kube-api-access-vz6r4\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.054380 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d17916ae-ef45-4999-8b6d-42f8476721e9-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.305934 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" event={"ID":"d17916ae-ef45-4999-8b6d-42f8476721e9","Type":"ContainerDied","Data":"401af2ff670333c550d972fd805237686195ed1354e47bd727544f98def4b907"} Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.305997 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="401af2ff670333c550d972fd805237686195ed1354e47bd727544f98def4b907" Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.306055 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj" Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.375093 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf"] Oct 02 08:15:04 crc kubenswrapper[4829]: I1002 08:15:04.386623 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323170-2tzgf"] Oct 02 08:15:05 crc kubenswrapper[4829]: I1002 08:15:05.484532 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3f20db-ce94-41ff-835f-86301d1ef391" path="/var/lib/kubelet/pods/7c3f20db-ce94-41ff-835f-86301d1ef391/volumes" Oct 02 08:15:25 crc kubenswrapper[4829]: I1002 08:15:25.329875 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:15:25 crc kubenswrapper[4829]: I1002 08:15:25.330541 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:15:55 crc kubenswrapper[4829]: I1002 08:15:55.329751 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:15:55 crc kubenswrapper[4829]: I1002 08:15:55.330525 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:15:59 crc kubenswrapper[4829]: I1002 08:15:59.817410 4829 scope.go:117] "RemoveContainer" containerID="927babadc432a63f7a214d1234b54b592230d6d84e17c703028640b6a553aeba" Oct 02 08:16:25 crc kubenswrapper[4829]: I1002 08:16:25.329167 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:16:25 crc kubenswrapper[4829]: I1002 08:16:25.330126 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:16:25 crc kubenswrapper[4829]: I1002 08:16:25.330203 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:16:25 crc kubenswrapper[4829]: I1002 08:16:25.331979 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:16:25 crc kubenswrapper[4829]: I1002 08:16:25.332089 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" gracePeriod=600 Oct 02 08:16:25 crc kubenswrapper[4829]: E1002 08:16:25.458226 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:16:26 crc kubenswrapper[4829]: I1002 08:16:26.398526 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" exitCode=0 Oct 02 08:16:26 crc kubenswrapper[4829]: I1002 08:16:26.398587 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3"} Oct 02 08:16:26 crc kubenswrapper[4829]: I1002 08:16:26.398657 4829 scope.go:117] "RemoveContainer" containerID="628e791b8bad71add3acfd79a863f9f6417cad671ec5f9feb48a872b7a63ccaa" Oct 02 08:16:26 crc kubenswrapper[4829]: I1002 08:16:26.400421 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:16:26 crc kubenswrapper[4829]: E1002 08:16:26.401408 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:16:38 crc kubenswrapper[4829]: I1002 08:16:38.461772 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:16:38 crc kubenswrapper[4829]: E1002 08:16:38.462850 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:16:53 crc kubenswrapper[4829]: I1002 08:16:53.462424 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:16:53 crc kubenswrapper[4829]: E1002 08:16:53.463521 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:17:05 crc kubenswrapper[4829]: I1002 08:17:05.461783 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:17:05 crc kubenswrapper[4829]: E1002 08:17:05.462993 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:17:20 crc kubenswrapper[4829]: I1002 08:17:20.461556 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:17:20 crc kubenswrapper[4829]: E1002 08:17:20.462384 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.600776 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n6tv6"] Oct 02 08:17:28 crc kubenswrapper[4829]: E1002 08:17:28.602708 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17916ae-ef45-4999-8b6d-42f8476721e9" containerName="collect-profiles" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.602727 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17916ae-ef45-4999-8b6d-42f8476721e9" containerName="collect-profiles" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.602964 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17916ae-ef45-4999-8b6d-42f8476721e9" containerName="collect-profiles" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.606559 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.636496 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n6tv6"] Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.791684 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-catalog-content\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.791990 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdlwp\" (UniqueName: \"kubernetes.io/projected/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-kube-api-access-kdlwp\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.792058 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-utilities\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.893944 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-utilities\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.894026 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-catalog-content\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.894114 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdlwp\" (UniqueName: \"kubernetes.io/projected/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-kube-api-access-kdlwp\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.894698 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-utilities\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.894729 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-catalog-content\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.913643 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdlwp\" (UniqueName: \"kubernetes.io/projected/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-kube-api-access-kdlwp\") pod \"community-operators-n6tv6\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:28 crc kubenswrapper[4829]: I1002 08:17:28.939261 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:29 crc kubenswrapper[4829]: I1002 08:17:29.457512 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n6tv6"] Oct 02 08:17:29 crc kubenswrapper[4829]: W1002 08:17:29.458457 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5cd28b0e_0797_4f16_8b34_53e00fbf4e3f.slice/crio-1831e29957f4dde26812cdfa6916904ec95b15d3b7a3c0f58c7e7cb76cdd0cf5 WatchSource:0}: Error finding container 1831e29957f4dde26812cdfa6916904ec95b15d3b7a3c0f58c7e7cb76cdd0cf5: Status 404 returned error can't find the container with id 1831e29957f4dde26812cdfa6916904ec95b15d3b7a3c0f58c7e7cb76cdd0cf5 Oct 02 08:17:30 crc kubenswrapper[4829]: I1002 08:17:30.157731 4829 generic.go:334] "Generic (PLEG): container finished" podID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerID="c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc" exitCode=0 Oct 02 08:17:30 crc kubenswrapper[4829]: I1002 08:17:30.157842 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerDied","Data":"c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc"} Oct 02 08:17:30 crc kubenswrapper[4829]: I1002 08:17:30.159288 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerStarted","Data":"1831e29957f4dde26812cdfa6916904ec95b15d3b7a3c0f58c7e7cb76cdd0cf5"} Oct 02 08:17:30 crc kubenswrapper[4829]: I1002 08:17:30.160971 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:17:31 crc kubenswrapper[4829]: I1002 08:17:31.170080 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerStarted","Data":"fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0"} Oct 02 08:17:31 crc kubenswrapper[4829]: I1002 08:17:31.461295 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:17:31 crc kubenswrapper[4829]: E1002 08:17:31.461548 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:17:32 crc kubenswrapper[4829]: I1002 08:17:32.182640 4829 generic.go:334] "Generic (PLEG): container finished" podID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerID="fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0" exitCode=0 Oct 02 08:17:32 crc kubenswrapper[4829]: I1002 08:17:32.182861 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerDied","Data":"fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0"} Oct 02 08:17:33 crc kubenswrapper[4829]: I1002 08:17:33.198969 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerStarted","Data":"d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02"} Oct 02 08:17:33 crc kubenswrapper[4829]: I1002 08:17:33.222985 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n6tv6" podStartSLOduration=2.737452345 podStartE2EDuration="5.222964524s" podCreationTimestamp="2025-10-02 08:17:28 +0000 UTC" firstStartedPulling="2025-10-02 08:17:30.160627966 +0000 UTC m=+3641.500276391" lastFinishedPulling="2025-10-02 08:17:32.646140155 +0000 UTC m=+3643.985788570" observedRunningTime="2025-10-02 08:17:33.220947061 +0000 UTC m=+3644.560595506" watchObservedRunningTime="2025-10-02 08:17:33.222964524 +0000 UTC m=+3644.562612939" Oct 02 08:17:38 crc kubenswrapper[4829]: I1002 08:17:38.940512 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:38 crc kubenswrapper[4829]: I1002 08:17:38.941276 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:39 crc kubenswrapper[4829]: I1002 08:17:39.020155 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:39 crc kubenswrapper[4829]: I1002 08:17:39.327185 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:39 crc kubenswrapper[4829]: I1002 08:17:39.384446 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n6tv6"] Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.276871 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n6tv6" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="registry-server" containerID="cri-o://d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02" gracePeriod=2 Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.872517 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.964149 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-catalog-content\") pod \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.964471 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-utilities\") pod \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.964590 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdlwp\" (UniqueName: \"kubernetes.io/projected/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-kube-api-access-kdlwp\") pod \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\" (UID: \"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f\") " Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.965497 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-utilities" (OuterVolumeSpecName: "utilities") pod "5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" (UID: "5cd28b0e-0797-4f16-8b34-53e00fbf4e3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.966852 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:17:41 crc kubenswrapper[4829]: I1002 08:17:41.973308 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-kube-api-access-kdlwp" (OuterVolumeSpecName: "kube-api-access-kdlwp") pod "5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" (UID: "5cd28b0e-0797-4f16-8b34-53e00fbf4e3f"). InnerVolumeSpecName "kube-api-access-kdlwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.013132 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" (UID: "5cd28b0e-0797-4f16-8b34-53e00fbf4e3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.068361 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdlwp\" (UniqueName: \"kubernetes.io/projected/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-kube-api-access-kdlwp\") on node \"crc\" DevicePath \"\"" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.068410 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.294056 4829 generic.go:334] "Generic (PLEG): container finished" podID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerID="d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02" exitCode=0 Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.294120 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerDied","Data":"d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02"} Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.294141 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n6tv6" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.294170 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n6tv6" event={"ID":"5cd28b0e-0797-4f16-8b34-53e00fbf4e3f","Type":"ContainerDied","Data":"1831e29957f4dde26812cdfa6916904ec95b15d3b7a3c0f58c7e7cb76cdd0cf5"} Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.294201 4829 scope.go:117] "RemoveContainer" containerID="d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.314490 4829 scope.go:117] "RemoveContainer" containerID="fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.349850 4829 scope.go:117] "RemoveContainer" containerID="c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.354653 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n6tv6"] Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.368971 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n6tv6"] Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.408714 4829 scope.go:117] "RemoveContainer" containerID="d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02" Oct 02 08:17:42 crc kubenswrapper[4829]: E1002 08:17:42.409077 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02\": container with ID starting with d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02 not found: ID does not exist" containerID="d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.409118 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02"} err="failed to get container status \"d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02\": rpc error: code = NotFound desc = could not find container \"d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02\": container with ID starting with d528e5139fe6e7ec012ceea5ae7163fb31589ef098f91984297c77ec1a813e02 not found: ID does not exist" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.409143 4829 scope.go:117] "RemoveContainer" containerID="fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0" Oct 02 08:17:42 crc kubenswrapper[4829]: E1002 08:17:42.409667 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0\": container with ID starting with fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0 not found: ID does not exist" containerID="fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.409696 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0"} err="failed to get container status \"fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0\": rpc error: code = NotFound desc = could not find container \"fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0\": container with ID starting with fc9e78a23f5698de55e9f7b3d625c641670bc66b9d477a2e79e3ccfa585ed8b0 not found: ID does not exist" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.409712 4829 scope.go:117] "RemoveContainer" containerID="c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc" Oct 02 08:17:42 crc kubenswrapper[4829]: E1002 08:17:42.409936 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc\": container with ID starting with c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc not found: ID does not exist" containerID="c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc" Oct 02 08:17:42 crc kubenswrapper[4829]: I1002 08:17:42.409960 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc"} err="failed to get container status \"c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc\": rpc error: code = NotFound desc = could not find container \"c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc\": container with ID starting with c2722a40ebe5fe69e8ce24b951f8f6d6beed44e8d0bd54e93ec0a93d88a7f6dc not found: ID does not exist" Oct 02 08:17:43 crc kubenswrapper[4829]: I1002 08:17:43.472509 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" path="/var/lib/kubelet/pods/5cd28b0e-0797-4f16-8b34-53e00fbf4e3f/volumes" Oct 02 08:17:46 crc kubenswrapper[4829]: I1002 08:17:46.460738 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:17:46 crc kubenswrapper[4829]: E1002 08:17:46.461415 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:18:01 crc kubenswrapper[4829]: I1002 08:18:01.460796 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:18:01 crc kubenswrapper[4829]: E1002 08:18:01.465535 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:18:16 crc kubenswrapper[4829]: I1002 08:18:16.461458 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:18:16 crc kubenswrapper[4829]: E1002 08:18:16.462729 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:18:30 crc kubenswrapper[4829]: I1002 08:18:30.461513 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:18:30 crc kubenswrapper[4829]: E1002 08:18:30.462547 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:18:41 crc kubenswrapper[4829]: I1002 08:18:41.461065 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:18:41 crc kubenswrapper[4829]: E1002 08:18:41.461796 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:18:52 crc kubenswrapper[4829]: I1002 08:18:52.461463 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:18:52 crc kubenswrapper[4829]: E1002 08:18:52.462620 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:19:04 crc kubenswrapper[4829]: I1002 08:19:04.461312 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:19:04 crc kubenswrapper[4829]: E1002 08:19:04.462965 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:19:18 crc kubenswrapper[4829]: I1002 08:19:18.462289 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:19:18 crc kubenswrapper[4829]: E1002 08:19:18.463343 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:19:31 crc kubenswrapper[4829]: I1002 08:19:31.461788 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:19:31 crc kubenswrapper[4829]: E1002 08:19:31.462664 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:19:45 crc kubenswrapper[4829]: I1002 08:19:45.461134 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:19:45 crc kubenswrapper[4829]: E1002 08:19:45.461909 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.561093 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ztwrp"] Oct 02 08:19:48 crc kubenswrapper[4829]: E1002 08:19:48.561841 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="registry-server" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.561854 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="registry-server" Oct 02 08:19:48 crc kubenswrapper[4829]: E1002 08:19:48.561870 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="extract-content" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.561876 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="extract-content" Oct 02 08:19:48 crc kubenswrapper[4829]: E1002 08:19:48.561904 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="extract-utilities" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.561910 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="extract-utilities" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.562146 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd28b0e-0797-4f16-8b34-53e00fbf4e3f" containerName="registry-server" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.564298 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.579627 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ztwrp"] Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.756137 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-utilities\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.756301 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf9dx\" (UniqueName: \"kubernetes.io/projected/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-kube-api-access-nf9dx\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.756382 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-catalog-content\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.858118 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf9dx\" (UniqueName: \"kubernetes.io/projected/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-kube-api-access-nf9dx\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.858212 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-catalog-content\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.858472 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-utilities\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.858808 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-catalog-content\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.859059 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-utilities\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.883596 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf9dx\" (UniqueName: \"kubernetes.io/projected/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-kube-api-access-nf9dx\") pod \"redhat-operators-ztwrp\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:48 crc kubenswrapper[4829]: I1002 08:19:48.923989 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:49 crc kubenswrapper[4829]: I1002 08:19:49.430912 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ztwrp"] Oct 02 08:19:49 crc kubenswrapper[4829]: I1002 08:19:49.772351 4829 generic.go:334] "Generic (PLEG): container finished" podID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerID="c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179" exitCode=0 Oct 02 08:19:49 crc kubenswrapper[4829]: I1002 08:19:49.772415 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerDied","Data":"c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179"} Oct 02 08:19:49 crc kubenswrapper[4829]: I1002 08:19:49.772648 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerStarted","Data":"d02885fe99305a4e426e76ee31ab4b411271545e82f83f87005565d72d005235"} Oct 02 08:19:51 crc kubenswrapper[4829]: I1002 08:19:51.806474 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerStarted","Data":"ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585"} Oct 02 08:19:51 crc kubenswrapper[4829]: I1002 08:19:51.925718 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xzrn2"] Oct 02 08:19:51 crc kubenswrapper[4829]: I1002 08:19:51.931403 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:51 crc kubenswrapper[4829]: I1002 08:19:51.938180 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzrn2"] Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.024953 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-catalog-content\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.025087 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zf4s5\" (UniqueName: \"kubernetes.io/projected/9d0814e5-b545-4140-9bc3-ba236ee38058-kube-api-access-zf4s5\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.025161 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-utilities\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.128466 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-catalog-content\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.128513 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-catalog-content\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.128745 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zf4s5\" (UniqueName: \"kubernetes.io/projected/9d0814e5-b545-4140-9bc3-ba236ee38058-kube-api-access-zf4s5\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.128900 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-utilities\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.129513 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-utilities\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.172288 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zf4s5\" (UniqueName: \"kubernetes.io/projected/9d0814e5-b545-4140-9bc3-ba236ee38058-kube-api-access-zf4s5\") pod \"certified-operators-xzrn2\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.270111 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:19:52 crc kubenswrapper[4829]: I1002 08:19:52.805825 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xzrn2"] Oct 02 08:19:52 crc kubenswrapper[4829]: W1002 08:19:52.809072 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d0814e5_b545_4140_9bc3_ba236ee38058.slice/crio-2ee75720cab529ebdb1193af9f01316edfd9b8910cf8fe9296f2fe3608ecb43b WatchSource:0}: Error finding container 2ee75720cab529ebdb1193af9f01316edfd9b8910cf8fe9296f2fe3608ecb43b: Status 404 returned error can't find the container with id 2ee75720cab529ebdb1193af9f01316edfd9b8910cf8fe9296f2fe3608ecb43b Oct 02 08:19:53 crc kubenswrapper[4829]: I1002 08:19:53.830697 4829 generic.go:334] "Generic (PLEG): container finished" podID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerID="64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd" exitCode=0 Oct 02 08:19:53 crc kubenswrapper[4829]: I1002 08:19:53.830863 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerDied","Data":"64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd"} Oct 02 08:19:53 crc kubenswrapper[4829]: I1002 08:19:53.831413 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerStarted","Data":"2ee75720cab529ebdb1193af9f01316edfd9b8910cf8fe9296f2fe3608ecb43b"} Oct 02 08:19:54 crc kubenswrapper[4829]: I1002 08:19:54.845962 4829 generic.go:334] "Generic (PLEG): container finished" podID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerID="ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585" exitCode=0 Oct 02 08:19:54 crc kubenswrapper[4829]: I1002 08:19:54.846087 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerDied","Data":"ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585"} Oct 02 08:19:54 crc kubenswrapper[4829]: I1002 08:19:54.851709 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerStarted","Data":"76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414"} Oct 02 08:19:55 crc kubenswrapper[4829]: I1002 08:19:55.868651 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerStarted","Data":"64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148"} Oct 02 08:19:55 crc kubenswrapper[4829]: I1002 08:19:55.892482 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ztwrp" podStartSLOduration=2.31663206 podStartE2EDuration="7.892459254s" podCreationTimestamp="2025-10-02 08:19:48 +0000 UTC" firstStartedPulling="2025-10-02 08:19:49.775713184 +0000 UTC m=+3781.115361589" lastFinishedPulling="2025-10-02 08:19:55.351540368 +0000 UTC m=+3786.691188783" observedRunningTime="2025-10-02 08:19:55.886827939 +0000 UTC m=+3787.226476394" watchObservedRunningTime="2025-10-02 08:19:55.892459254 +0000 UTC m=+3787.232107699" Oct 02 08:19:56 crc kubenswrapper[4829]: I1002 08:19:56.881038 4829 generic.go:334] "Generic (PLEG): container finished" podID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerID="76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414" exitCode=0 Oct 02 08:19:56 crc kubenswrapper[4829]: I1002 08:19:56.881383 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerDied","Data":"76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414"} Oct 02 08:19:57 crc kubenswrapper[4829]: I1002 08:19:57.896414 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerStarted","Data":"31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7"} Oct 02 08:19:57 crc kubenswrapper[4829]: I1002 08:19:57.918435 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xzrn2" podStartSLOduration=3.42210061 podStartE2EDuration="6.918420379s" podCreationTimestamp="2025-10-02 08:19:51 +0000 UTC" firstStartedPulling="2025-10-02 08:19:53.833588718 +0000 UTC m=+3785.173237163" lastFinishedPulling="2025-10-02 08:19:57.329908527 +0000 UTC m=+3788.669556932" observedRunningTime="2025-10-02 08:19:57.917216191 +0000 UTC m=+3789.256864596" watchObservedRunningTime="2025-10-02 08:19:57.918420379 +0000 UTC m=+3789.258068784" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.124133 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4wq"] Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.126532 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.140208 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4wq"] Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.265336 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-catalog-content\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.265864 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-utilities\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.265950 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84hp6\" (UniqueName: \"kubernetes.io/projected/80fbbfed-2111-443a-8e3d-ce5b13a516e8-kube-api-access-84hp6\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.367435 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84hp6\" (UniqueName: \"kubernetes.io/projected/80fbbfed-2111-443a-8e3d-ce5b13a516e8-kube-api-access-84hp6\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.367486 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-catalog-content\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.367740 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-utilities\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.368149 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-catalog-content\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.368282 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-utilities\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.390368 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84hp6\" (UniqueName: \"kubernetes.io/projected/80fbbfed-2111-443a-8e3d-ce5b13a516e8-kube-api-access-84hp6\") pod \"redhat-marketplace-gp4wq\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.442574 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.461388 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:19:58 crc kubenswrapper[4829]: E1002 08:19:58.461661 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.918339 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4wq"] Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.924219 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:58 crc kubenswrapper[4829]: I1002 08:19:58.924276 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:19:59 crc kubenswrapper[4829]: I1002 08:19:59.917235 4829 generic.go:334] "Generic (PLEG): container finished" podID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerID="f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085" exitCode=0 Oct 02 08:19:59 crc kubenswrapper[4829]: I1002 08:19:59.917290 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4wq" event={"ID":"80fbbfed-2111-443a-8e3d-ce5b13a516e8","Type":"ContainerDied","Data":"f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085"} Oct 02 08:19:59 crc kubenswrapper[4829]: I1002 08:19:59.917338 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4wq" event={"ID":"80fbbfed-2111-443a-8e3d-ce5b13a516e8","Type":"ContainerStarted","Data":"d3e69ec32ff66740b7f0ea69df7ebbcd893b3b3dc4639755048db43b682e5cef"} Oct 02 08:19:59 crc kubenswrapper[4829]: I1002 08:19:59.996530 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ztwrp" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="registry-server" probeResult="failure" output=< Oct 02 08:19:59 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:19:59 crc kubenswrapper[4829]: > Oct 02 08:20:01 crc kubenswrapper[4829]: I1002 08:20:01.941839 4829 generic.go:334] "Generic (PLEG): container finished" podID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerID="04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a" exitCode=0 Oct 02 08:20:01 crc kubenswrapper[4829]: I1002 08:20:01.942371 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4wq" event={"ID":"80fbbfed-2111-443a-8e3d-ce5b13a516e8","Type":"ContainerDied","Data":"04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a"} Oct 02 08:20:02 crc kubenswrapper[4829]: I1002 08:20:02.271064 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:20:02 crc kubenswrapper[4829]: I1002 08:20:02.271390 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:20:02 crc kubenswrapper[4829]: I1002 08:20:02.320146 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:20:02 crc kubenswrapper[4829]: I1002 08:20:02.952682 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4wq" event={"ID":"80fbbfed-2111-443a-8e3d-ce5b13a516e8","Type":"ContainerStarted","Data":"e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386"} Oct 02 08:20:03 crc kubenswrapper[4829]: I1002 08:20:03.021239 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:20:03 crc kubenswrapper[4829]: I1002 08:20:03.040970 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gp4wq" podStartSLOduration=2.521599085 podStartE2EDuration="5.040947553s" podCreationTimestamp="2025-10-02 08:19:58 +0000 UTC" firstStartedPulling="2025-10-02 08:19:59.919607406 +0000 UTC m=+3791.259255811" lastFinishedPulling="2025-10-02 08:20:02.438955874 +0000 UTC m=+3793.778604279" observedRunningTime="2025-10-02 08:20:02.975938287 +0000 UTC m=+3794.315586712" watchObservedRunningTime="2025-10-02 08:20:03.040947553 +0000 UTC m=+3794.380595958" Oct 02 08:20:04 crc kubenswrapper[4829]: I1002 08:20:04.712859 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzrn2"] Oct 02 08:20:04 crc kubenswrapper[4829]: I1002 08:20:04.971107 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xzrn2" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="registry-server" containerID="cri-o://31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7" gracePeriod=2 Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.584009 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.719066 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-utilities\") pod \"9d0814e5-b545-4140-9bc3-ba236ee38058\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.719130 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zf4s5\" (UniqueName: \"kubernetes.io/projected/9d0814e5-b545-4140-9bc3-ba236ee38058-kube-api-access-zf4s5\") pod \"9d0814e5-b545-4140-9bc3-ba236ee38058\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.719289 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-catalog-content\") pod \"9d0814e5-b545-4140-9bc3-ba236ee38058\" (UID: \"9d0814e5-b545-4140-9bc3-ba236ee38058\") " Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.719871 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-utilities" (OuterVolumeSpecName: "utilities") pod "9d0814e5-b545-4140-9bc3-ba236ee38058" (UID: "9d0814e5-b545-4140-9bc3-ba236ee38058"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.724452 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d0814e5-b545-4140-9bc3-ba236ee38058-kube-api-access-zf4s5" (OuterVolumeSpecName: "kube-api-access-zf4s5") pod "9d0814e5-b545-4140-9bc3-ba236ee38058" (UID: "9d0814e5-b545-4140-9bc3-ba236ee38058"). InnerVolumeSpecName "kube-api-access-zf4s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.760420 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d0814e5-b545-4140-9bc3-ba236ee38058" (UID: "9d0814e5-b545-4140-9bc3-ba236ee38058"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.821843 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.821882 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zf4s5\" (UniqueName: \"kubernetes.io/projected/9d0814e5-b545-4140-9bc3-ba236ee38058-kube-api-access-zf4s5\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.821896 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d0814e5-b545-4140-9bc3-ba236ee38058-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.982994 4829 generic.go:334] "Generic (PLEG): container finished" podID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerID="31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7" exitCode=0 Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.983045 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerDied","Data":"31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7"} Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.983083 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xzrn2" event={"ID":"9d0814e5-b545-4140-9bc3-ba236ee38058","Type":"ContainerDied","Data":"2ee75720cab529ebdb1193af9f01316edfd9b8910cf8fe9296f2fe3608ecb43b"} Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.983110 4829 scope.go:117] "RemoveContainer" containerID="31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7" Oct 02 08:20:05 crc kubenswrapper[4829]: I1002 08:20:05.983104 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xzrn2" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.017204 4829 scope.go:117] "RemoveContainer" containerID="76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.033743 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xzrn2"] Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.051976 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xzrn2"] Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.060554 4829 scope.go:117] "RemoveContainer" containerID="64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.108484 4829 scope.go:117] "RemoveContainer" containerID="31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7" Oct 02 08:20:06 crc kubenswrapper[4829]: E1002 08:20:06.109346 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7\": container with ID starting with 31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7 not found: ID does not exist" containerID="31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.109408 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7"} err="failed to get container status \"31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7\": rpc error: code = NotFound desc = could not find container \"31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7\": container with ID starting with 31aeeb7ce7ac580db5f85d6065458887ea6e1ee294ae07ad77e55853fea672b7 not found: ID does not exist" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.109452 4829 scope.go:117] "RemoveContainer" containerID="76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414" Oct 02 08:20:06 crc kubenswrapper[4829]: E1002 08:20:06.110074 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414\": container with ID starting with 76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414 not found: ID does not exist" containerID="76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.110133 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414"} err="failed to get container status \"76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414\": rpc error: code = NotFound desc = could not find container \"76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414\": container with ID starting with 76f504d44de36f6a47b9cd0a2eb81ec26a0e87816043443db0eddcb628bd4414 not found: ID does not exist" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.110173 4829 scope.go:117] "RemoveContainer" containerID="64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd" Oct 02 08:20:06 crc kubenswrapper[4829]: E1002 08:20:06.110600 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd\": container with ID starting with 64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd not found: ID does not exist" containerID="64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd" Oct 02 08:20:06 crc kubenswrapper[4829]: I1002 08:20:06.110659 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd"} err="failed to get container status \"64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd\": rpc error: code = NotFound desc = could not find container \"64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd\": container with ID starting with 64e681af1cbf79a5b7715c348413ec7f8b4b8f1505aa8ba65d275f19b81640dd not found: ID does not exist" Oct 02 08:20:07 crc kubenswrapper[4829]: I1002 08:20:07.476149 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" path="/var/lib/kubelet/pods/9d0814e5-b545-4140-9bc3-ba236ee38058/volumes" Oct 02 08:20:08 crc kubenswrapper[4829]: I1002 08:20:08.443449 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:20:08 crc kubenswrapper[4829]: I1002 08:20:08.443974 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:20:08 crc kubenswrapper[4829]: I1002 08:20:08.531406 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:20:09 crc kubenswrapper[4829]: I1002 08:20:09.059748 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:20:09 crc kubenswrapper[4829]: I1002 08:20:09.476198 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:20:09 crc kubenswrapper[4829]: E1002 08:20:09.476837 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:20:09 crc kubenswrapper[4829]: I1002 08:20:09.522925 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4wq"] Oct 02 08:20:10 crc kubenswrapper[4829]: I1002 08:20:10.008545 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-ztwrp" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="registry-server" probeResult="failure" output=< Oct 02 08:20:10 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:20:10 crc kubenswrapper[4829]: > Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.031514 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gp4wq" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="registry-server" containerID="cri-o://e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386" gracePeriod=2 Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.567737 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.660882 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-catalog-content\") pod \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.661046 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-utilities\") pod \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.661104 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84hp6\" (UniqueName: \"kubernetes.io/projected/80fbbfed-2111-443a-8e3d-ce5b13a516e8-kube-api-access-84hp6\") pod \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\" (UID: \"80fbbfed-2111-443a-8e3d-ce5b13a516e8\") " Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.661913 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-utilities" (OuterVolumeSpecName: "utilities") pod "80fbbfed-2111-443a-8e3d-ce5b13a516e8" (UID: "80fbbfed-2111-443a-8e3d-ce5b13a516e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.671590 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80fbbfed-2111-443a-8e3d-ce5b13a516e8-kube-api-access-84hp6" (OuterVolumeSpecName: "kube-api-access-84hp6") pod "80fbbfed-2111-443a-8e3d-ce5b13a516e8" (UID: "80fbbfed-2111-443a-8e3d-ce5b13a516e8"). InnerVolumeSpecName "kube-api-access-84hp6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.678877 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "80fbbfed-2111-443a-8e3d-ce5b13a516e8" (UID: "80fbbfed-2111-443a-8e3d-ce5b13a516e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.762485 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.762512 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/80fbbfed-2111-443a-8e3d-ce5b13a516e8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:11 crc kubenswrapper[4829]: I1002 08:20:11.762522 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84hp6\" (UniqueName: \"kubernetes.io/projected/80fbbfed-2111-443a-8e3d-ce5b13a516e8-kube-api-access-84hp6\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.047144 4829 generic.go:334] "Generic (PLEG): container finished" podID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerID="e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386" exitCode=0 Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.047185 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4wq" event={"ID":"80fbbfed-2111-443a-8e3d-ce5b13a516e8","Type":"ContainerDied","Data":"e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386"} Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.047214 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gp4wq" event={"ID":"80fbbfed-2111-443a-8e3d-ce5b13a516e8","Type":"ContainerDied","Data":"d3e69ec32ff66740b7f0ea69df7ebbcd893b3b3dc4639755048db43b682e5cef"} Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.047252 4829 scope.go:117] "RemoveContainer" containerID="e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.047277 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gp4wq" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.077679 4829 scope.go:117] "RemoveContainer" containerID="04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.114399 4829 scope.go:117] "RemoveContainer" containerID="f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.116676 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4wq"] Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.128989 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gp4wq"] Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.187784 4829 scope.go:117] "RemoveContainer" containerID="e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386" Oct 02 08:20:12 crc kubenswrapper[4829]: E1002 08:20:12.188314 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386\": container with ID starting with e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386 not found: ID does not exist" containerID="e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.188403 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386"} err="failed to get container status \"e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386\": rpc error: code = NotFound desc = could not find container \"e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386\": container with ID starting with e99dc71caab7b2c7812b06200278e7e114003b0f7cfd4a4a812432fd134bd386 not found: ID does not exist" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.188438 4829 scope.go:117] "RemoveContainer" containerID="04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a" Oct 02 08:20:12 crc kubenswrapper[4829]: E1002 08:20:12.188798 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a\": container with ID starting with 04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a not found: ID does not exist" containerID="04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.188853 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a"} err="failed to get container status \"04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a\": rpc error: code = NotFound desc = could not find container \"04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a\": container with ID starting with 04e64bc84c05225f95f8d62cc6e5ba97cf5c1974e62c4ea235d425ecb4c74f3a not found: ID does not exist" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.188886 4829 scope.go:117] "RemoveContainer" containerID="f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085" Oct 02 08:20:12 crc kubenswrapper[4829]: E1002 08:20:12.189175 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085\": container with ID starting with f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085 not found: ID does not exist" containerID="f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085" Oct 02 08:20:12 crc kubenswrapper[4829]: I1002 08:20:12.189253 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085"} err="failed to get container status \"f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085\": rpc error: code = NotFound desc = could not find container \"f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085\": container with ID starting with f125eadb38fdd3fd57290c81f78f7c19c6b27b5bf345cf0445cc99cf1a9d0085 not found: ID does not exist" Oct 02 08:20:13 crc kubenswrapper[4829]: I1002 08:20:13.482311 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" path="/var/lib/kubelet/pods/80fbbfed-2111-443a-8e3d-ce5b13a516e8/volumes" Oct 02 08:20:18 crc kubenswrapper[4829]: I1002 08:20:18.993562 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:20:19 crc kubenswrapper[4829]: I1002 08:20:19.063641 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:20:19 crc kubenswrapper[4829]: I1002 08:20:19.750610 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ztwrp"] Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.156418 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ztwrp" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="registry-server" containerID="cri-o://64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148" gracePeriod=2 Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.462124 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:20:20 crc kubenswrapper[4829]: E1002 08:20:20.462745 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.715320 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.763671 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf9dx\" (UniqueName: \"kubernetes.io/projected/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-kube-api-access-nf9dx\") pod \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.763811 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-catalog-content\") pod \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.763889 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-utilities\") pod \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\" (UID: \"76d7354c-e0b4-425f-afd0-6b3c9d17dec7\") " Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.766700 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-utilities" (OuterVolumeSpecName: "utilities") pod "76d7354c-e0b4-425f-afd0-6b3c9d17dec7" (UID: "76d7354c-e0b4-425f-afd0-6b3c9d17dec7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.770915 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-kube-api-access-nf9dx" (OuterVolumeSpecName: "kube-api-access-nf9dx") pod "76d7354c-e0b4-425f-afd0-6b3c9d17dec7" (UID: "76d7354c-e0b4-425f-afd0-6b3c9d17dec7"). InnerVolumeSpecName "kube-api-access-nf9dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.866190 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.866600 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf9dx\" (UniqueName: \"kubernetes.io/projected/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-kube-api-access-nf9dx\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.869975 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76d7354c-e0b4-425f-afd0-6b3c9d17dec7" (UID: "76d7354c-e0b4-425f-afd0-6b3c9d17dec7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:20:20 crc kubenswrapper[4829]: I1002 08:20:20.968624 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76d7354c-e0b4-425f-afd0-6b3c9d17dec7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.172330 4829 generic.go:334] "Generic (PLEG): container finished" podID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerID="64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148" exitCode=0 Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.172398 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerDied","Data":"64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148"} Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.172438 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ztwrp" event={"ID":"76d7354c-e0b4-425f-afd0-6b3c9d17dec7","Type":"ContainerDied","Data":"d02885fe99305a4e426e76ee31ab4b411271545e82f83f87005565d72d005235"} Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.172468 4829 scope.go:117] "RemoveContainer" containerID="64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.172545 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ztwrp" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.215862 4829 scope.go:117] "RemoveContainer" containerID="ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.231818 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ztwrp"] Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.243964 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ztwrp"] Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.250908 4829 scope.go:117] "RemoveContainer" containerID="c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.306041 4829 scope.go:117] "RemoveContainer" containerID="64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148" Oct 02 08:20:21 crc kubenswrapper[4829]: E1002 08:20:21.306991 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148\": container with ID starting with 64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148 not found: ID does not exist" containerID="64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.307034 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148"} err="failed to get container status \"64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148\": rpc error: code = NotFound desc = could not find container \"64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148\": container with ID starting with 64450a4af14bc64f9bd42009099d30cc1a4240f262e1ff15b602ae964a7d0148 not found: ID does not exist" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.307061 4829 scope.go:117] "RemoveContainer" containerID="ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585" Oct 02 08:20:21 crc kubenswrapper[4829]: E1002 08:20:21.307486 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585\": container with ID starting with ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585 not found: ID does not exist" containerID="ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.307520 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585"} err="failed to get container status \"ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585\": rpc error: code = NotFound desc = could not find container \"ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585\": container with ID starting with ab192fc21e34c3e57755544635eff1c4336296e22eb30ab1edf81f0577a91585 not found: ID does not exist" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.307541 4829 scope.go:117] "RemoveContainer" containerID="c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179" Oct 02 08:20:21 crc kubenswrapper[4829]: E1002 08:20:21.307897 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179\": container with ID starting with c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179 not found: ID does not exist" containerID="c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.307957 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179"} err="failed to get container status \"c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179\": rpc error: code = NotFound desc = could not find container \"c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179\": container with ID starting with c8acfa7d4d3152ee68da3336abc530ae91af7fac6a6b8318aacb3353e05cd179 not found: ID does not exist" Oct 02 08:20:21 crc kubenswrapper[4829]: I1002 08:20:21.481407 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" path="/var/lib/kubelet/pods/76d7354c-e0b4-425f-afd0-6b3c9d17dec7/volumes" Oct 02 08:20:33 crc kubenswrapper[4829]: I1002 08:20:33.461490 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:20:33 crc kubenswrapper[4829]: E1002 08:20:33.462487 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:20:44 crc kubenswrapper[4829]: I1002 08:20:44.461277 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:20:44 crc kubenswrapper[4829]: E1002 08:20:44.462313 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:20:57 crc kubenswrapper[4829]: I1002 08:20:57.461880 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:20:57 crc kubenswrapper[4829]: E1002 08:20:57.462923 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:21:08 crc kubenswrapper[4829]: I1002 08:21:08.461607 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:21:08 crc kubenswrapper[4829]: E1002 08:21:08.462434 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:21:23 crc kubenswrapper[4829]: I1002 08:21:23.460266 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:21:23 crc kubenswrapper[4829]: E1002 08:21:23.461050 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:21:37 crc kubenswrapper[4829]: I1002 08:21:37.461682 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:21:38 crc kubenswrapper[4829]: I1002 08:21:38.097604 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"2683adbd31b9a9a40c760a813137428dd153676318c00d6863967e5711740fb7"} Oct 02 08:22:32 crc kubenswrapper[4829]: E1002 08:22:32.717310 4829 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:50866->38.102.83.200:43623: write tcp 38.102.83.200:50866->38.102.83.200:43623: write: broken pipe Oct 02 08:23:55 crc kubenswrapper[4829]: I1002 08:23:55.328989 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:23:55 crc kubenswrapper[4829]: I1002 08:23:55.329792 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:24:25 crc kubenswrapper[4829]: I1002 08:24:25.329286 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:24:25 crc kubenswrapper[4829]: I1002 08:24:25.329947 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.329573 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.330181 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.330273 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.331401 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2683adbd31b9a9a40c760a813137428dd153676318c00d6863967e5711740fb7"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.331510 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://2683adbd31b9a9a40c760a813137428dd153676318c00d6863967e5711740fb7" gracePeriod=600 Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.530963 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"2683adbd31b9a9a40c760a813137428dd153676318c00d6863967e5711740fb7"} Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.531356 4829 scope.go:117] "RemoveContainer" containerID="f8adc303d88e767798e06f0a3c0365f486d6b1d9155ac73af9ecff991b05a9f3" Oct 02 08:24:55 crc kubenswrapper[4829]: I1002 08:24:55.530965 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="2683adbd31b9a9a40c760a813137428dd153676318c00d6863967e5711740fb7" exitCode=0 Oct 02 08:24:56 crc kubenswrapper[4829]: I1002 08:24:56.543963 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859"} Oct 02 08:26:55 crc kubenswrapper[4829]: I1002 08:26:55.329355 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:26:55 crc kubenswrapper[4829]: I1002 08:26:55.329849 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:27:25 crc kubenswrapper[4829]: I1002 08:27:25.328938 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:27:25 crc kubenswrapper[4829]: I1002 08:27:25.329729 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.474766 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jbkk2"] Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.475953 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="extract-utilities" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.475980 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="extract-utilities" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.475999 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="extract-content" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476012 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="extract-content" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476033 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="extract-utilities" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476047 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="extract-utilities" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476102 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476114 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476133 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="extract-content" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476145 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="extract-content" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476167 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476180 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476209 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="extract-content" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476221 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="extract-content" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476292 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476310 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: E1002 08:27:30.476330 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="extract-utilities" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476344 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="extract-utilities" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476761 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d0814e5-b545-4140-9bc3-ba236ee38058" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476796 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="80fbbfed-2111-443a-8e3d-ce5b13a516e8" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.476822 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="76d7354c-e0b4-425f-afd0-6b3c9d17dec7" containerName="registry-server" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.479609 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.495635 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbkk2"] Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.648366 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5tnz\" (UniqueName: \"kubernetes.io/projected/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-kube-api-access-p5tnz\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.648416 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-catalog-content\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.648440 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-utilities\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.750024 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5tnz\" (UniqueName: \"kubernetes.io/projected/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-kube-api-access-p5tnz\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.750318 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-catalog-content\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.750343 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-utilities\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.750800 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-utilities\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.750922 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-catalog-content\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.787589 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5tnz\" (UniqueName: \"kubernetes.io/projected/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-kube-api-access-p5tnz\") pod \"community-operators-jbkk2\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:30 crc kubenswrapper[4829]: I1002 08:27:30.813020 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:31 crc kubenswrapper[4829]: I1002 08:27:31.330856 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbkk2"] Oct 02 08:27:31 crc kubenswrapper[4829]: I1002 08:27:31.432619 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerStarted","Data":"9c9c3dc51b809e4cc8f2b0511d216b6fbea22d5de9965841a348cff49eb11877"} Oct 02 08:27:32 crc kubenswrapper[4829]: I1002 08:27:32.446640 4829 generic.go:334] "Generic (PLEG): container finished" podID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerID="176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75" exitCode=0 Oct 02 08:27:32 crc kubenswrapper[4829]: I1002 08:27:32.446755 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerDied","Data":"176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75"} Oct 02 08:27:32 crc kubenswrapper[4829]: I1002 08:27:32.449352 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:27:34 crc kubenswrapper[4829]: I1002 08:27:34.473303 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerStarted","Data":"467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94"} Oct 02 08:27:35 crc kubenswrapper[4829]: I1002 08:27:35.485914 4829 generic.go:334] "Generic (PLEG): container finished" podID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerID="467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94" exitCode=0 Oct 02 08:27:35 crc kubenswrapper[4829]: I1002 08:27:35.485984 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerDied","Data":"467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94"} Oct 02 08:27:36 crc kubenswrapper[4829]: I1002 08:27:36.497667 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerStarted","Data":"af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e"} Oct 02 08:27:36 crc kubenswrapper[4829]: I1002 08:27:36.527721 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jbkk2" podStartSLOduration=3.047405823 podStartE2EDuration="6.527700878s" podCreationTimestamp="2025-10-02 08:27:30 +0000 UTC" firstStartedPulling="2025-10-02 08:27:32.449052326 +0000 UTC m=+4243.788700741" lastFinishedPulling="2025-10-02 08:27:35.929347351 +0000 UTC m=+4247.268995796" observedRunningTime="2025-10-02 08:27:36.51882555 +0000 UTC m=+4247.858473955" watchObservedRunningTime="2025-10-02 08:27:36.527700878 +0000 UTC m=+4247.867349273" Oct 02 08:27:40 crc kubenswrapper[4829]: I1002 08:27:40.813494 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:40 crc kubenswrapper[4829]: I1002 08:27:40.813962 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:40 crc kubenswrapper[4829]: I1002 08:27:40.901131 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:41 crc kubenswrapper[4829]: I1002 08:27:41.644809 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:41 crc kubenswrapper[4829]: I1002 08:27:41.704088 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbkk2"] Oct 02 08:27:43 crc kubenswrapper[4829]: I1002 08:27:43.583527 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jbkk2" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="registry-server" containerID="cri-o://af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e" gracePeriod=2 Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.245594 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.364444 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-catalog-content\") pod \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.364962 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-utilities\") pod \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.365050 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5tnz\" (UniqueName: \"kubernetes.io/projected/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-kube-api-access-p5tnz\") pod \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\" (UID: \"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78\") " Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.367045 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-utilities" (OuterVolumeSpecName: "utilities") pod "fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" (UID: "fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.382469 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-kube-api-access-p5tnz" (OuterVolumeSpecName: "kube-api-access-p5tnz") pod "fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" (UID: "fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78"). InnerVolumeSpecName "kube-api-access-p5tnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.422490 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" (UID: "fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.466532 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.466564 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5tnz\" (UniqueName: \"kubernetes.io/projected/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-kube-api-access-p5tnz\") on node \"crc\" DevicePath \"\"" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.466575 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.600017 4829 generic.go:334] "Generic (PLEG): container finished" podID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerID="af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e" exitCode=0 Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.600105 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerDied","Data":"af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e"} Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.600175 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbkk2" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.600213 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbkk2" event={"ID":"fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78","Type":"ContainerDied","Data":"9c9c3dc51b809e4cc8f2b0511d216b6fbea22d5de9965841a348cff49eb11877"} Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.600269 4829 scope.go:117] "RemoveContainer" containerID="af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.629431 4829 scope.go:117] "RemoveContainer" containerID="467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.642527 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbkk2"] Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.651209 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jbkk2"] Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.685323 4829 scope.go:117] "RemoveContainer" containerID="176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.729451 4829 scope.go:117] "RemoveContainer" containerID="af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e" Oct 02 08:27:44 crc kubenswrapper[4829]: E1002 08:27:44.732584 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e\": container with ID starting with af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e not found: ID does not exist" containerID="af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.732634 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e"} err="failed to get container status \"af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e\": rpc error: code = NotFound desc = could not find container \"af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e\": container with ID starting with af3f5f8162cf19554f0ace77f661763d4a5d778715a4493c16ba4f4b3862503e not found: ID does not exist" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.732668 4829 scope.go:117] "RemoveContainer" containerID="467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94" Oct 02 08:27:44 crc kubenswrapper[4829]: E1002 08:27:44.733494 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94\": container with ID starting with 467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94 not found: ID does not exist" containerID="467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.733550 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94"} err="failed to get container status \"467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94\": rpc error: code = NotFound desc = could not find container \"467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94\": container with ID starting with 467b73797a14b4c92782305823f5aef1f69b3e7304658b3169e903d2545afd94 not found: ID does not exist" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.733588 4829 scope.go:117] "RemoveContainer" containerID="176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75" Oct 02 08:27:44 crc kubenswrapper[4829]: E1002 08:27:44.734105 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75\": container with ID starting with 176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75 not found: ID does not exist" containerID="176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75" Oct 02 08:27:44 crc kubenswrapper[4829]: I1002 08:27:44.734140 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75"} err="failed to get container status \"176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75\": rpc error: code = NotFound desc = could not find container \"176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75\": container with ID starting with 176b5bc4b23b81bd4efd62b48677a6e42f242b10d2a7f42f2a4c146c49049f75 not found: ID does not exist" Oct 02 08:27:45 crc kubenswrapper[4829]: I1002 08:27:45.476723 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" path="/var/lib/kubelet/pods/fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78/volumes" Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.328882 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.330423 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.330531 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.331367 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.331506 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" gracePeriod=600 Oct 02 08:27:55 crc kubenswrapper[4829]: E1002 08:27:55.459260 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.722916 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" exitCode=0 Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.722968 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859"} Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.723044 4829 scope.go:117] "RemoveContainer" containerID="2683adbd31b9a9a40c760a813137428dd153676318c00d6863967e5711740fb7" Oct 02 08:27:55 crc kubenswrapper[4829]: I1002 08:27:55.723841 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:27:55 crc kubenswrapper[4829]: E1002 08:27:55.724448 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:28:09 crc kubenswrapper[4829]: I1002 08:28:09.474948 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:28:09 crc kubenswrapper[4829]: E1002 08:28:09.476325 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:28:24 crc kubenswrapper[4829]: I1002 08:28:24.460934 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:28:24 crc kubenswrapper[4829]: E1002 08:28:24.462197 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:28:35 crc kubenswrapper[4829]: I1002 08:28:35.461885 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:28:35 crc kubenswrapper[4829]: E1002 08:28:35.462952 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:28:48 crc kubenswrapper[4829]: I1002 08:28:48.461920 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:28:48 crc kubenswrapper[4829]: E1002 08:28:48.463133 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:28:59 crc kubenswrapper[4829]: I1002 08:28:59.476150 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:28:59 crc kubenswrapper[4829]: E1002 08:28:59.477348 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:29:11 crc kubenswrapper[4829]: I1002 08:29:11.461149 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:29:11 crc kubenswrapper[4829]: E1002 08:29:11.464122 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:29:24 crc kubenswrapper[4829]: I1002 08:29:24.461489 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:29:24 crc kubenswrapper[4829]: E1002 08:29:24.462330 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:29:36 crc kubenswrapper[4829]: I1002 08:29:36.461579 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:29:36 crc kubenswrapper[4829]: E1002 08:29:36.462898 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:29:47 crc kubenswrapper[4829]: I1002 08:29:47.461219 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:29:47 crc kubenswrapper[4829]: E1002 08:29:47.462330 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:29:59 crc kubenswrapper[4829]: I1002 08:29:59.470792 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:29:59 crc kubenswrapper[4829]: E1002 08:29:59.471668 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.179866 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm"] Oct 02 08:30:00 crc kubenswrapper[4829]: E1002 08:30:00.182459 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="registry-server" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.182506 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="registry-server" Oct 02 08:30:00 crc kubenswrapper[4829]: E1002 08:30:00.182589 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="extract-utilities" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.182609 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="extract-utilities" Oct 02 08:30:00 crc kubenswrapper[4829]: E1002 08:30:00.182654 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="extract-content" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.182670 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="extract-content" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.183806 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb0bbeb3-e4c5-453b-b7d0-ee1aa8440c78" containerName="registry-server" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.187820 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.191944 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.192844 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.232734 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm"] Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.325563 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b60d086-b6f7-4baa-92c6-945a916abed6-secret-volume\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.325639 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzbq9\" (UniqueName: \"kubernetes.io/projected/5b60d086-b6f7-4baa-92c6-945a916abed6-kube-api-access-bzbq9\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.325689 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b60d086-b6f7-4baa-92c6-945a916abed6-config-volume\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.427008 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b60d086-b6f7-4baa-92c6-945a916abed6-secret-volume\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.427068 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzbq9\" (UniqueName: \"kubernetes.io/projected/5b60d086-b6f7-4baa-92c6-945a916abed6-kube-api-access-bzbq9\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.427119 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b60d086-b6f7-4baa-92c6-945a916abed6-config-volume\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.427969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b60d086-b6f7-4baa-92c6-945a916abed6-config-volume\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.442835 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b60d086-b6f7-4baa-92c6-945a916abed6-secret-volume\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.445886 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzbq9\" (UniqueName: \"kubernetes.io/projected/5b60d086-b6f7-4baa-92c6-945a916abed6-kube-api-access-bzbq9\") pod \"collect-profiles-29323230-hpddm\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:00 crc kubenswrapper[4829]: I1002 08:30:00.528162 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:01 crc kubenswrapper[4829]: I1002 08:30:01.059583 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm"] Oct 02 08:30:01 crc kubenswrapper[4829]: I1002 08:30:01.206897 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" event={"ID":"5b60d086-b6f7-4baa-92c6-945a916abed6","Type":"ContainerStarted","Data":"a07f34955d737ccacf0475658f4d3c6cf07a8f03cd3a64ae9d18be2f2cf06ea3"} Oct 02 08:30:02 crc kubenswrapper[4829]: I1002 08:30:02.221828 4829 generic.go:334] "Generic (PLEG): container finished" podID="5b60d086-b6f7-4baa-92c6-945a916abed6" containerID="e10fec59d0e1a753ec9e13bd65014e0993ead613e148aebaaaea7655dc36a952" exitCode=0 Oct 02 08:30:02 crc kubenswrapper[4829]: I1002 08:30:02.222120 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" event={"ID":"5b60d086-b6f7-4baa-92c6-945a916abed6","Type":"ContainerDied","Data":"e10fec59d0e1a753ec9e13bd65014e0993ead613e148aebaaaea7655dc36a952"} Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.789055 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.895758 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzbq9\" (UniqueName: \"kubernetes.io/projected/5b60d086-b6f7-4baa-92c6-945a916abed6-kube-api-access-bzbq9\") pod \"5b60d086-b6f7-4baa-92c6-945a916abed6\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.895857 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b60d086-b6f7-4baa-92c6-945a916abed6-secret-volume\") pod \"5b60d086-b6f7-4baa-92c6-945a916abed6\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.895928 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b60d086-b6f7-4baa-92c6-945a916abed6-config-volume\") pod \"5b60d086-b6f7-4baa-92c6-945a916abed6\" (UID: \"5b60d086-b6f7-4baa-92c6-945a916abed6\") " Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.896575 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b60d086-b6f7-4baa-92c6-945a916abed6-config-volume" (OuterVolumeSpecName: "config-volume") pod "5b60d086-b6f7-4baa-92c6-945a916abed6" (UID: "5b60d086-b6f7-4baa-92c6-945a916abed6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.901934 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b60d086-b6f7-4baa-92c6-945a916abed6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5b60d086-b6f7-4baa-92c6-945a916abed6" (UID: "5b60d086-b6f7-4baa-92c6-945a916abed6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.902289 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b60d086-b6f7-4baa-92c6-945a916abed6-kube-api-access-bzbq9" (OuterVolumeSpecName: "kube-api-access-bzbq9") pod "5b60d086-b6f7-4baa-92c6-945a916abed6" (UID: "5b60d086-b6f7-4baa-92c6-945a916abed6"). InnerVolumeSpecName "kube-api-access-bzbq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.998408 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzbq9\" (UniqueName: \"kubernetes.io/projected/5b60d086-b6f7-4baa-92c6-945a916abed6-kube-api-access-bzbq9\") on node \"crc\" DevicePath \"\"" Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.998652 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b60d086-b6f7-4baa-92c6-945a916abed6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:30:03 crc kubenswrapper[4829]: I1002 08:30:03.998663 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b60d086-b6f7-4baa-92c6-945a916abed6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:30:04 crc kubenswrapper[4829]: I1002 08:30:04.250699 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" event={"ID":"5b60d086-b6f7-4baa-92c6-945a916abed6","Type":"ContainerDied","Data":"a07f34955d737ccacf0475658f4d3c6cf07a8f03cd3a64ae9d18be2f2cf06ea3"} Oct 02 08:30:04 crc kubenswrapper[4829]: I1002 08:30:04.251072 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a07f34955d737ccacf0475658f4d3c6cf07a8f03cd3a64ae9d18be2f2cf06ea3" Oct 02 08:30:04 crc kubenswrapper[4829]: I1002 08:30:04.250793 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323230-hpddm" Oct 02 08:30:04 crc kubenswrapper[4829]: I1002 08:30:04.881713 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw"] Oct 02 08:30:04 crc kubenswrapper[4829]: I1002 08:30:04.895341 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323185-6hnnw"] Oct 02 08:30:05 crc kubenswrapper[4829]: I1002 08:30:05.483149 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b5d0a66-00ea-4f01-9040-6b53de05c254" path="/var/lib/kubelet/pods/9b5d0a66-00ea-4f01-9040-6b53de05c254/volumes" Oct 02 08:30:13 crc kubenswrapper[4829]: I1002 08:30:13.461990 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:30:13 crc kubenswrapper[4829]: E1002 08:30:13.463216 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:30:24 crc kubenswrapper[4829]: I1002 08:30:24.461770 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:30:24 crc kubenswrapper[4829]: E1002 08:30:24.462656 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:30:38 crc kubenswrapper[4829]: I1002 08:30:38.462297 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:30:38 crc kubenswrapper[4829]: E1002 08:30:38.463323 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.461642 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:30:52 crc kubenswrapper[4829]: E1002 08:30:52.462633 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.665560 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-62744"] Oct 02 08:30:52 crc kubenswrapper[4829]: E1002 08:30:52.666363 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b60d086-b6f7-4baa-92c6-945a916abed6" containerName="collect-profiles" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.666395 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b60d086-b6f7-4baa-92c6-945a916abed6" containerName="collect-profiles" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.666934 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b60d086-b6f7-4baa-92c6-945a916abed6" containerName="collect-profiles" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.694980 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.705746 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62744"] Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.847259 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-catalog-content\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.847559 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9mrg\" (UniqueName: \"kubernetes.io/projected/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-kube-api-access-s9mrg\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.847682 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-utilities\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.950410 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-catalog-content\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.950859 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9mrg\" (UniqueName: \"kubernetes.io/projected/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-kube-api-access-s9mrg\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.951106 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-utilities\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.950903 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-catalog-content\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.951453 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-utilities\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:52 crc kubenswrapper[4829]: I1002 08:30:52.987283 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9mrg\" (UniqueName: \"kubernetes.io/projected/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-kube-api-access-s9mrg\") pod \"certified-operators-62744\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:53 crc kubenswrapper[4829]: I1002 08:30:53.018023 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:30:53 crc kubenswrapper[4829]: I1002 08:30:53.529989 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62744"] Oct 02 08:30:53 crc kubenswrapper[4829]: I1002 08:30:53.812631 4829 generic.go:334] "Generic (PLEG): container finished" podID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerID="2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d" exitCode=0 Oct 02 08:30:53 crc kubenswrapper[4829]: I1002 08:30:53.812690 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerDied","Data":"2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d"} Oct 02 08:30:53 crc kubenswrapper[4829]: I1002 08:30:53.812909 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerStarted","Data":"ee12bcd00f5b7b8984efae24ea0a9af9375ae7cc17546a02ae92b131380901b0"} Oct 02 08:30:55 crc kubenswrapper[4829]: I1002 08:30:55.839309 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerStarted","Data":"ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d"} Oct 02 08:30:56 crc kubenswrapper[4829]: I1002 08:30:56.852547 4829 generic.go:334] "Generic (PLEG): container finished" podID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerID="ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d" exitCode=0 Oct 02 08:30:56 crc kubenswrapper[4829]: I1002 08:30:56.852679 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerDied","Data":"ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d"} Oct 02 08:30:57 crc kubenswrapper[4829]: I1002 08:30:57.866905 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerStarted","Data":"826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151"} Oct 02 08:30:57 crc kubenswrapper[4829]: I1002 08:30:57.900875 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-62744" podStartSLOduration=2.259257831 podStartE2EDuration="5.900846753s" podCreationTimestamp="2025-10-02 08:30:52 +0000 UTC" firstStartedPulling="2025-10-02 08:30:53.814173081 +0000 UTC m=+4445.153821476" lastFinishedPulling="2025-10-02 08:30:57.455761983 +0000 UTC m=+4448.795410398" observedRunningTime="2025-10-02 08:30:57.890731128 +0000 UTC m=+4449.230379563" watchObservedRunningTime="2025-10-02 08:30:57.900846753 +0000 UTC m=+4449.240495178" Oct 02 08:31:00 crc kubenswrapper[4829]: I1002 08:31:00.342360 4829 scope.go:117] "RemoveContainer" containerID="198a1730723e0f0dfd27f4ff4b99c4bac6cc7699d97cabad8e606d84ffeb7de4" Oct 02 08:31:03 crc kubenswrapper[4829]: I1002 08:31:03.018216 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:31:03 crc kubenswrapper[4829]: I1002 08:31:03.018831 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:31:03 crc kubenswrapper[4829]: I1002 08:31:03.063654 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:31:03 crc kubenswrapper[4829]: I1002 08:31:03.461822 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:31:03 crc kubenswrapper[4829]: E1002 08:31:03.462691 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:31:04 crc kubenswrapper[4829]: I1002 08:31:04.003959 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:31:04 crc kubenswrapper[4829]: I1002 08:31:04.052957 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62744"] Oct 02 08:31:05 crc kubenswrapper[4829]: I1002 08:31:05.953530 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-62744" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="registry-server" containerID="cri-o://826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151" gracePeriod=2 Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.605794 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.773107 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9mrg\" (UniqueName: \"kubernetes.io/projected/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-kube-api-access-s9mrg\") pod \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.773202 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-utilities\") pod \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.773371 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-catalog-content\") pod \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\" (UID: \"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48\") " Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.774633 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-utilities" (OuterVolumeSpecName: "utilities") pod "b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" (UID: "b1c6ffee-8728-4b8e-8150-c7c6aafa3b48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.786037 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-kube-api-access-s9mrg" (OuterVolumeSpecName: "kube-api-access-s9mrg") pod "b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" (UID: "b1c6ffee-8728-4b8e-8150-c7c6aafa3b48"). InnerVolumeSpecName "kube-api-access-s9mrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.824266 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" (UID: "b1c6ffee-8728-4b8e-8150-c7c6aafa3b48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.875881 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.875927 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9mrg\" (UniqueName: \"kubernetes.io/projected/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-kube-api-access-s9mrg\") on node \"crc\" DevicePath \"\"" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.875940 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.970279 4829 generic.go:334] "Generic (PLEG): container finished" podID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerID="826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151" exitCode=0 Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.970353 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerDied","Data":"826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151"} Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.970352 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62744" Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.970417 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62744" event={"ID":"b1c6ffee-8728-4b8e-8150-c7c6aafa3b48","Type":"ContainerDied","Data":"ee12bcd00f5b7b8984efae24ea0a9af9375ae7cc17546a02ae92b131380901b0"} Oct 02 08:31:06 crc kubenswrapper[4829]: I1002 08:31:06.970447 4829 scope.go:117] "RemoveContainer" containerID="826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.017750 4829 scope.go:117] "RemoveContainer" containerID="ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.021354 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62744"] Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.031890 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-62744"] Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.046337 4829 scope.go:117] "RemoveContainer" containerID="2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.099793 4829 scope.go:117] "RemoveContainer" containerID="826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151" Oct 02 08:31:07 crc kubenswrapper[4829]: E1002 08:31:07.100401 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151\": container with ID starting with 826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151 not found: ID does not exist" containerID="826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.100474 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151"} err="failed to get container status \"826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151\": rpc error: code = NotFound desc = could not find container \"826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151\": container with ID starting with 826c08b6be24871f326184dc4c34f72613dade003b1739ec50a5513e77753151 not found: ID does not exist" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.100515 4829 scope.go:117] "RemoveContainer" containerID="ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d" Oct 02 08:31:07 crc kubenswrapper[4829]: E1002 08:31:07.100854 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d\": container with ID starting with ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d not found: ID does not exist" containerID="ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.100890 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d"} err="failed to get container status \"ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d\": rpc error: code = NotFound desc = could not find container \"ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d\": container with ID starting with ac6e3851e7f2d6fda5330fc81e1d4a4f0f2381c472b0e56d277ca16f0cf4e78d not found: ID does not exist" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.100919 4829 scope.go:117] "RemoveContainer" containerID="2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d" Oct 02 08:31:07 crc kubenswrapper[4829]: E1002 08:31:07.101270 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d\": container with ID starting with 2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d not found: ID does not exist" containerID="2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.101311 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d"} err="failed to get container status \"2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d\": rpc error: code = NotFound desc = could not find container \"2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d\": container with ID starting with 2940560095dbb6faad69acf6fdd616565c3475b22beebe968d37301b46cff46d not found: ID does not exist" Oct 02 08:31:07 crc kubenswrapper[4829]: I1002 08:31:07.473981 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" path="/var/lib/kubelet/pods/b1c6ffee-8728-4b8e-8150-c7c6aafa3b48/volumes" Oct 02 08:31:18 crc kubenswrapper[4829]: I1002 08:31:18.461654 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:31:18 crc kubenswrapper[4829]: E1002 08:31:18.462560 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:31:30 crc kubenswrapper[4829]: I1002 08:31:30.461958 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:31:30 crc kubenswrapper[4829]: E1002 08:31:30.462728 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.093893 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4wgzr"] Oct 02 08:31:40 crc kubenswrapper[4829]: E1002 08:31:40.095261 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="extract-utilities" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.095286 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="extract-utilities" Oct 02 08:31:40 crc kubenswrapper[4829]: E1002 08:31:40.095307 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="registry-server" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.095319 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="registry-server" Oct 02 08:31:40 crc kubenswrapper[4829]: E1002 08:31:40.095378 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="extract-content" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.095393 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="extract-content" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.095780 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c6ffee-8728-4b8e-8150-c7c6aafa3b48" containerName="registry-server" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.098439 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.111369 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wgzr"] Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.292786 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97x5f\" (UniqueName: \"kubernetes.io/projected/4ba9bfa3-0b75-4005-9158-555ab46f6764-kube-api-access-97x5f\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.293273 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-catalog-content\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.293365 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-utilities\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.395496 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-utilities\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.395777 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97x5f\" (UniqueName: \"kubernetes.io/projected/4ba9bfa3-0b75-4005-9158-555ab46f6764-kube-api-access-97x5f\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.395860 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-catalog-content\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.396078 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-utilities\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.396352 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-catalog-content\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.422297 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97x5f\" (UniqueName: \"kubernetes.io/projected/4ba9bfa3-0b75-4005-9158-555ab46f6764-kube-api-access-97x5f\") pod \"redhat-operators-4wgzr\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.434746 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:40 crc kubenswrapper[4829]: I1002 08:31:40.931478 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4wgzr"] Oct 02 08:31:41 crc kubenswrapper[4829]: I1002 08:31:41.429680 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerID="6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d" exitCode=0 Oct 02 08:31:41 crc kubenswrapper[4829]: I1002 08:31:41.429784 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerDied","Data":"6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d"} Oct 02 08:31:41 crc kubenswrapper[4829]: I1002 08:31:41.430277 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerStarted","Data":"a880a0b1ae1720bb7a354c8515dd611665e8bafbacb5a7c7a9e74a57ef5ef8f8"} Oct 02 08:31:42 crc kubenswrapper[4829]: I1002 08:31:42.460868 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:31:42 crc kubenswrapper[4829]: E1002 08:31:42.461561 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:31:43 crc kubenswrapper[4829]: I1002 08:31:43.479731 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerStarted","Data":"ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff"} Oct 02 08:31:46 crc kubenswrapper[4829]: I1002 08:31:46.515274 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerID="ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff" exitCode=0 Oct 02 08:31:46 crc kubenswrapper[4829]: I1002 08:31:46.515361 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerDied","Data":"ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff"} Oct 02 08:31:47 crc kubenswrapper[4829]: I1002 08:31:47.534360 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerStarted","Data":"11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc"} Oct 02 08:31:47 crc kubenswrapper[4829]: I1002 08:31:47.563386 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4wgzr" podStartSLOduration=1.860428779 podStartE2EDuration="7.56335301s" podCreationTimestamp="2025-10-02 08:31:40 +0000 UTC" firstStartedPulling="2025-10-02 08:31:41.432407262 +0000 UTC m=+4492.772055667" lastFinishedPulling="2025-10-02 08:31:47.135331443 +0000 UTC m=+4498.474979898" observedRunningTime="2025-10-02 08:31:47.552854823 +0000 UTC m=+4498.892503238" watchObservedRunningTime="2025-10-02 08:31:47.56335301 +0000 UTC m=+4498.903001455" Oct 02 08:31:50 crc kubenswrapper[4829]: I1002 08:31:50.435074 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:50 crc kubenswrapper[4829]: I1002 08:31:50.435427 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:31:51 crc kubenswrapper[4829]: I1002 08:31:51.508611 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4wgzr" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="registry-server" probeResult="failure" output=< Oct 02 08:31:51 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:31:51 crc kubenswrapper[4829]: > Oct 02 08:31:55 crc kubenswrapper[4829]: I1002 08:31:55.460725 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:31:55 crc kubenswrapper[4829]: E1002 08:31:55.461845 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:32:01 crc kubenswrapper[4829]: I1002 08:32:01.494311 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4wgzr" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="registry-server" probeResult="failure" output=< Oct 02 08:32:01 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:32:01 crc kubenswrapper[4829]: > Oct 02 08:32:09 crc kubenswrapper[4829]: I1002 08:32:09.477706 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:32:09 crc kubenswrapper[4829]: E1002 08:32:09.479154 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:32:10 crc kubenswrapper[4829]: I1002 08:32:10.524269 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:32:10 crc kubenswrapper[4829]: I1002 08:32:10.596346 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:32:11 crc kubenswrapper[4829]: I1002 08:32:11.283888 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wgzr"] Oct 02 08:32:11 crc kubenswrapper[4829]: I1002 08:32:11.799774 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4wgzr" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="registry-server" containerID="cri-o://11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc" gracePeriod=2 Oct 02 08:32:12 crc kubenswrapper[4829]: E1002 08:32:12.132951 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ba9bfa3_0b75_4005_9158_555ab46f6764.slice/crio-conmon-11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ba9bfa3_0b75_4005_9158_555ab46f6764.slice/crio-11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc.scope\": RecentStats: unable to find data in memory cache]" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.314945 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.394723 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-utilities\") pod \"4ba9bfa3-0b75-4005-9158-555ab46f6764\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.395062 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97x5f\" (UniqueName: \"kubernetes.io/projected/4ba9bfa3-0b75-4005-9158-555ab46f6764-kube-api-access-97x5f\") pod \"4ba9bfa3-0b75-4005-9158-555ab46f6764\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.395112 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-catalog-content\") pod \"4ba9bfa3-0b75-4005-9158-555ab46f6764\" (UID: \"4ba9bfa3-0b75-4005-9158-555ab46f6764\") " Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.395840 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-utilities" (OuterVolumeSpecName: "utilities") pod "4ba9bfa3-0b75-4005-9158-555ab46f6764" (UID: "4ba9bfa3-0b75-4005-9158-555ab46f6764"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.401388 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ba9bfa3-0b75-4005-9158-555ab46f6764-kube-api-access-97x5f" (OuterVolumeSpecName: "kube-api-access-97x5f") pod "4ba9bfa3-0b75-4005-9158-555ab46f6764" (UID: "4ba9bfa3-0b75-4005-9158-555ab46f6764"). InnerVolumeSpecName "kube-api-access-97x5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.493997 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ba9bfa3-0b75-4005-9158-555ab46f6764" (UID: "4ba9bfa3-0b75-4005-9158-555ab46f6764"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.497797 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97x5f\" (UniqueName: \"kubernetes.io/projected/4ba9bfa3-0b75-4005-9158-555ab46f6764-kube-api-access-97x5f\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.497910 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.497966 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ba9bfa3-0b75-4005-9158-555ab46f6764-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.817210 4829 generic.go:334] "Generic (PLEG): container finished" podID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerID="11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc" exitCode=0 Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.817292 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerDied","Data":"11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc"} Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.817325 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4wgzr" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.817340 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4wgzr" event={"ID":"4ba9bfa3-0b75-4005-9158-555ab46f6764","Type":"ContainerDied","Data":"a880a0b1ae1720bb7a354c8515dd611665e8bafbacb5a7c7a9e74a57ef5ef8f8"} Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.817372 4829 scope.go:117] "RemoveContainer" containerID="11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.859978 4829 scope.go:117] "RemoveContainer" containerID="ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.866435 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4wgzr"] Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.875589 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4wgzr"] Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.909756 4829 scope.go:117] "RemoveContainer" containerID="6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.960052 4829 scope.go:117] "RemoveContainer" containerID="11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc" Oct 02 08:32:12 crc kubenswrapper[4829]: E1002 08:32:12.960736 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc\": container with ID starting with 11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc not found: ID does not exist" containerID="11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.960801 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc"} err="failed to get container status \"11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc\": rpc error: code = NotFound desc = could not find container \"11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc\": container with ID starting with 11f58242f8d8294e359ab065ba2c5e2fb96a7f69de341bbf4b65079f149094cc not found: ID does not exist" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.960838 4829 scope.go:117] "RemoveContainer" containerID="ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff" Oct 02 08:32:12 crc kubenswrapper[4829]: E1002 08:32:12.961365 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff\": container with ID starting with ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff not found: ID does not exist" containerID="ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.961407 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff"} err="failed to get container status \"ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff\": rpc error: code = NotFound desc = could not find container \"ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff\": container with ID starting with ea80599cf92d7adea136a13f6b7911eedc2ebd9f4c66233ff747a9e78e8a26ff not found: ID does not exist" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.961436 4829 scope.go:117] "RemoveContainer" containerID="6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d" Oct 02 08:32:12 crc kubenswrapper[4829]: E1002 08:32:12.961902 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d\": container with ID starting with 6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d not found: ID does not exist" containerID="6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d" Oct 02 08:32:12 crc kubenswrapper[4829]: I1002 08:32:12.961950 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d"} err="failed to get container status \"6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d\": rpc error: code = NotFound desc = could not find container \"6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d\": container with ID starting with 6643d5289db5588e2673d321b3a4fdcc0f690c62c9b0be3972debd1e5e58e92d not found: ID does not exist" Oct 02 08:32:13 crc kubenswrapper[4829]: I1002 08:32:13.483165 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" path="/var/lib/kubelet/pods/4ba9bfa3-0b75-4005-9158-555ab46f6764/volumes" Oct 02 08:32:24 crc kubenswrapper[4829]: I1002 08:32:24.460855 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:32:24 crc kubenswrapper[4829]: E1002 08:32:24.461650 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:32:39 crc kubenswrapper[4829]: I1002 08:32:39.476824 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:32:39 crc kubenswrapper[4829]: E1002 08:32:39.477746 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.645375 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-57cw4"] Oct 02 08:32:44 crc kubenswrapper[4829]: E1002 08:32:44.648120 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="extract-utilities" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.648286 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="extract-utilities" Oct 02 08:32:44 crc kubenswrapper[4829]: E1002 08:32:44.648429 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="registry-server" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.648530 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="registry-server" Oct 02 08:32:44 crc kubenswrapper[4829]: E1002 08:32:44.648618 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="extract-content" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.648713 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="extract-content" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.649149 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ba9bfa3-0b75-4005-9158-555ab46f6764" containerName="registry-server" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.651568 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.672016 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-57cw4"] Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.782489 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-catalog-content\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.782549 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvk5d\" (UniqueName: \"kubernetes.io/projected/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-kube-api-access-zvk5d\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.782619 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-utilities\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.883920 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-catalog-content\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.884143 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvk5d\" (UniqueName: \"kubernetes.io/projected/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-kube-api-access-zvk5d\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.884275 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-utilities\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.884415 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-catalog-content\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.884713 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-utilities\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:44 crc kubenswrapper[4829]: I1002 08:32:44.993515 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvk5d\" (UniqueName: \"kubernetes.io/projected/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-kube-api-access-zvk5d\") pod \"redhat-marketplace-57cw4\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:45 crc kubenswrapper[4829]: I1002 08:32:45.279856 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:45 crc kubenswrapper[4829]: I1002 08:32:45.814685 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-57cw4"] Oct 02 08:32:46 crc kubenswrapper[4829]: I1002 08:32:46.224359 4829 generic.go:334] "Generic (PLEG): container finished" podID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerID="c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6" exitCode=0 Oct 02 08:32:46 crc kubenswrapper[4829]: I1002 08:32:46.224445 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerDied","Data":"c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6"} Oct 02 08:32:46 crc kubenswrapper[4829]: I1002 08:32:46.224760 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerStarted","Data":"c397464544c586b0223d65aadd84e2592765fde1f38fc80bc969b39917b0d32b"} Oct 02 08:32:46 crc kubenswrapper[4829]: I1002 08:32:46.227569 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:32:47 crc kubenswrapper[4829]: I1002 08:32:47.240269 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerStarted","Data":"a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804"} Oct 02 08:32:48 crc kubenswrapper[4829]: I1002 08:32:48.252249 4829 generic.go:334] "Generic (PLEG): container finished" podID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerID="a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804" exitCode=0 Oct 02 08:32:48 crc kubenswrapper[4829]: I1002 08:32:48.252362 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerDied","Data":"a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804"} Oct 02 08:32:49 crc kubenswrapper[4829]: I1002 08:32:49.265884 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerStarted","Data":"09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6"} Oct 02 08:32:49 crc kubenswrapper[4829]: I1002 08:32:49.289542 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-57cw4" podStartSLOduration=2.551033295 podStartE2EDuration="5.289523132s" podCreationTimestamp="2025-10-02 08:32:44 +0000 UTC" firstStartedPulling="2025-10-02 08:32:46.227024516 +0000 UTC m=+4557.566672951" lastFinishedPulling="2025-10-02 08:32:48.965514363 +0000 UTC m=+4560.305162788" observedRunningTime="2025-10-02 08:32:49.28015553 +0000 UTC m=+4560.619803955" watchObservedRunningTime="2025-10-02 08:32:49.289523132 +0000 UTC m=+4560.629171537" Oct 02 08:32:52 crc kubenswrapper[4829]: I1002 08:32:52.460764 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:32:52 crc kubenswrapper[4829]: E1002 08:32:52.461511 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:32:55 crc kubenswrapper[4829]: I1002 08:32:55.280699 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:55 crc kubenswrapper[4829]: I1002 08:32:55.281365 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:55 crc kubenswrapper[4829]: I1002 08:32:55.362440 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:55 crc kubenswrapper[4829]: I1002 08:32:55.422340 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:55 crc kubenswrapper[4829]: I1002 08:32:55.611161 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-57cw4"] Oct 02 08:32:57 crc kubenswrapper[4829]: I1002 08:32:57.363499 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-57cw4" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="registry-server" containerID="cri-o://09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6" gracePeriod=2 Oct 02 08:32:57 crc kubenswrapper[4829]: I1002 08:32:57.981397 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.083306 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-utilities\") pod \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.083472 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-catalog-content\") pod \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.083545 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvk5d\" (UniqueName: \"kubernetes.io/projected/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-kube-api-access-zvk5d\") pod \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\" (UID: \"0ab99da6-53fd-4968-9f37-337f7dd0a7b5\") " Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.085904 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-utilities" (OuterVolumeSpecName: "utilities") pod "0ab99da6-53fd-4968-9f37-337f7dd0a7b5" (UID: "0ab99da6-53fd-4968-9f37-337f7dd0a7b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.098274 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ab99da6-53fd-4968-9f37-337f7dd0a7b5" (UID: "0ab99da6-53fd-4968-9f37-337f7dd0a7b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.185895 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.185934 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.378958 4829 generic.go:334] "Generic (PLEG): container finished" podID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerID="09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6" exitCode=0 Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.379034 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-57cw4" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.379038 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerDied","Data":"09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6"} Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.381366 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-57cw4" event={"ID":"0ab99da6-53fd-4968-9f37-337f7dd0a7b5","Type":"ContainerDied","Data":"c397464544c586b0223d65aadd84e2592765fde1f38fc80bc969b39917b0d32b"} Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.381391 4829 scope.go:117] "RemoveContainer" containerID="09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.413933 4829 scope.go:117] "RemoveContainer" containerID="a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.496431 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-kube-api-access-zvk5d" (OuterVolumeSpecName: "kube-api-access-zvk5d") pod "0ab99da6-53fd-4968-9f37-337f7dd0a7b5" (UID: "0ab99da6-53fd-4968-9f37-337f7dd0a7b5"). InnerVolumeSpecName "kube-api-access-zvk5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.511962 4829 scope.go:117] "RemoveContainer" containerID="c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.595636 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvk5d\" (UniqueName: \"kubernetes.io/projected/0ab99da6-53fd-4968-9f37-337f7dd0a7b5-kube-api-access-zvk5d\") on node \"crc\" DevicePath \"\"" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.752903 4829 scope.go:117] "RemoveContainer" containerID="09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6" Oct 02 08:32:58 crc kubenswrapper[4829]: E1002 08:32:58.753658 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6\": container with ID starting with 09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6 not found: ID does not exist" containerID="09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.753696 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6"} err="failed to get container status \"09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6\": rpc error: code = NotFound desc = could not find container \"09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6\": container with ID starting with 09a41befa584d8e8bd1fac9e82d06923380d1fb55c711bf0f78b1c9847cf86b6 not found: ID does not exist" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.753721 4829 scope.go:117] "RemoveContainer" containerID="a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804" Oct 02 08:32:58 crc kubenswrapper[4829]: E1002 08:32:58.754383 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804\": container with ID starting with a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804 not found: ID does not exist" containerID="a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.754453 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804"} err="failed to get container status \"a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804\": rpc error: code = NotFound desc = could not find container \"a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804\": container with ID starting with a0fe29cde754380f1abf70018dcf1ff0036846501845ec9f9bd81fbd7e8d5804 not found: ID does not exist" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.754496 4829 scope.go:117] "RemoveContainer" containerID="c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6" Oct 02 08:32:58 crc kubenswrapper[4829]: E1002 08:32:58.755523 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6\": container with ID starting with c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6 not found: ID does not exist" containerID="c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.755568 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6"} err="failed to get container status \"c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6\": rpc error: code = NotFound desc = could not find container \"c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6\": container with ID starting with c9659d282e59828b2372d78f254b1f3dcdf5a3322eff43ea5ca999efd9cf91d6 not found: ID does not exist" Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.797094 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-57cw4"] Oct 02 08:32:58 crc kubenswrapper[4829]: I1002 08:32:58.808316 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-57cw4"] Oct 02 08:32:59 crc kubenswrapper[4829]: I1002 08:32:59.482575 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" path="/var/lib/kubelet/pods/0ab99da6-53fd-4968-9f37-337f7dd0a7b5/volumes" Oct 02 08:33:06 crc kubenswrapper[4829]: I1002 08:33:06.461306 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:33:07 crc kubenswrapper[4829]: I1002 08:33:07.494213 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"d9daaaad0373f37594d146a03b1ae22f94640ac1896cf827746fd46a113423ab"} Oct 02 08:33:20 crc kubenswrapper[4829]: E1002 08:33:20.015330 4829 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:39296->38.102.83.200:43623: write tcp 38.102.83.200:39296->38.102.83.200:43623: write: broken pipe Oct 02 08:35:25 crc kubenswrapper[4829]: I1002 08:35:25.329166 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:35:25 crc kubenswrapper[4829]: I1002 08:35:25.329688 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:35:55 crc kubenswrapper[4829]: I1002 08:35:55.329715 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:35:55 crc kubenswrapper[4829]: I1002 08:35:55.330343 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.329916 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.332405 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.332698 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.334135 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d9daaaad0373f37594d146a03b1ae22f94640ac1896cf827746fd46a113423ab"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.334524 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://d9daaaad0373f37594d146a03b1ae22f94640ac1896cf827746fd46a113423ab" gracePeriod=600 Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.945774 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="d9daaaad0373f37594d146a03b1ae22f94640ac1896cf827746fd46a113423ab" exitCode=0 Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.945815 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"d9daaaad0373f37594d146a03b1ae22f94640ac1896cf827746fd46a113423ab"} Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.946070 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5"} Oct 02 08:36:25 crc kubenswrapper[4829]: I1002 08:36:25.946094 4829 scope.go:117] "RemoveContainer" containerID="d5fabab850132a8f822601a0d966dd8262884c5985f3903313d41a1317f2e859" Oct 02 08:37:35 crc kubenswrapper[4829]: E1002 08:37:35.326057 4829 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:37428->38.102.83.200:43623: write tcp 38.102.83.200:37428->38.102.83.200:43623: write: broken pipe Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.329613 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.330372 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.378282 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cr597"] Oct 02 08:38:25 crc kubenswrapper[4829]: E1002 08:38:25.378724 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="extract-content" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.378745 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="extract-content" Oct 02 08:38:25 crc kubenswrapper[4829]: E1002 08:38:25.378762 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="extract-utilities" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.378770 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="extract-utilities" Oct 02 08:38:25 crc kubenswrapper[4829]: E1002 08:38:25.378797 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="registry-server" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.378806 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="registry-server" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.379027 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ab99da6-53fd-4968-9f37-337f7dd0a7b5" containerName="registry-server" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.380735 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.397253 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cr597"] Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.496090 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67z44\" (UniqueName: \"kubernetes.io/projected/01715007-c25e-42a2-a7f5-fea2131d154c-kube-api-access-67z44\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.496607 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-utilities\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.497349 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-catalog-content\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.599501 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67z44\" (UniqueName: \"kubernetes.io/projected/01715007-c25e-42a2-a7f5-fea2131d154c-kube-api-access-67z44\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.599565 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-utilities\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.599623 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-catalog-content\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.600053 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-catalog-content\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.600559 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-utilities\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.619498 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67z44\" (UniqueName: \"kubernetes.io/projected/01715007-c25e-42a2-a7f5-fea2131d154c-kube-api-access-67z44\") pod \"community-operators-cr597\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:25 crc kubenswrapper[4829]: I1002 08:38:25.712091 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:26 crc kubenswrapper[4829]: I1002 08:38:26.302599 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cr597"] Oct 02 08:38:26 crc kubenswrapper[4829]: W1002 08:38:26.308047 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01715007_c25e_42a2_a7f5_fea2131d154c.slice/crio-8b458169847cbaf35aad300f418d9325a9f4c0b66c9c46a1b0338de235834a1b WatchSource:0}: Error finding container 8b458169847cbaf35aad300f418d9325a9f4c0b66c9c46a1b0338de235834a1b: Status 404 returned error can't find the container with id 8b458169847cbaf35aad300f418d9325a9f4c0b66c9c46a1b0338de235834a1b Oct 02 08:38:26 crc kubenswrapper[4829]: I1002 08:38:26.436264 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerStarted","Data":"8b458169847cbaf35aad300f418d9325a9f4c0b66c9c46a1b0338de235834a1b"} Oct 02 08:38:27 crc kubenswrapper[4829]: I1002 08:38:27.450780 4829 generic.go:334] "Generic (PLEG): container finished" podID="01715007-c25e-42a2-a7f5-fea2131d154c" containerID="0b5a54e38e8450a389d4f16b45b2e9c05d10cd3f2917f1a6f2e678ceab48253e" exitCode=0 Oct 02 08:38:27 crc kubenswrapper[4829]: I1002 08:38:27.451094 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerDied","Data":"0b5a54e38e8450a389d4f16b45b2e9c05d10cd3f2917f1a6f2e678ceab48253e"} Oct 02 08:38:27 crc kubenswrapper[4829]: I1002 08:38:27.455989 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:38:28 crc kubenswrapper[4829]: I1002 08:38:28.465853 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerStarted","Data":"27fc2a7d7934d2615b7815106b6b67fabd606269d639ff709656e39e47d802c9"} Oct 02 08:38:30 crc kubenswrapper[4829]: I1002 08:38:30.494205 4829 generic.go:334] "Generic (PLEG): container finished" podID="01715007-c25e-42a2-a7f5-fea2131d154c" containerID="27fc2a7d7934d2615b7815106b6b67fabd606269d639ff709656e39e47d802c9" exitCode=0 Oct 02 08:38:30 crc kubenswrapper[4829]: I1002 08:38:30.494419 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerDied","Data":"27fc2a7d7934d2615b7815106b6b67fabd606269d639ff709656e39e47d802c9"} Oct 02 08:38:31 crc kubenswrapper[4829]: I1002 08:38:31.510711 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerStarted","Data":"5dab16f693141f9da639930da2bd902921b2215db0bd5052ef801cf4ec8eec67"} Oct 02 08:38:31 crc kubenswrapper[4829]: I1002 08:38:31.533857 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cr597" podStartSLOduration=3.067906852 podStartE2EDuration="6.533840008s" podCreationTimestamp="2025-10-02 08:38:25 +0000 UTC" firstStartedPulling="2025-10-02 08:38:27.455586728 +0000 UTC m=+4898.795235163" lastFinishedPulling="2025-10-02 08:38:30.921519884 +0000 UTC m=+4902.261168319" observedRunningTime="2025-10-02 08:38:31.528136709 +0000 UTC m=+4902.867785124" watchObservedRunningTime="2025-10-02 08:38:31.533840008 +0000 UTC m=+4902.873488423" Oct 02 08:38:35 crc kubenswrapper[4829]: I1002 08:38:35.713456 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:35 crc kubenswrapper[4829]: I1002 08:38:35.715672 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:35 crc kubenswrapper[4829]: I1002 08:38:35.771634 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:36 crc kubenswrapper[4829]: I1002 08:38:36.653905 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:36 crc kubenswrapper[4829]: I1002 08:38:36.715965 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cr597"] Oct 02 08:38:38 crc kubenswrapper[4829]: I1002 08:38:38.606742 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cr597" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="registry-server" containerID="cri-o://5dab16f693141f9da639930da2bd902921b2215db0bd5052ef801cf4ec8eec67" gracePeriod=2 Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.624934 4829 generic.go:334] "Generic (PLEG): container finished" podID="01715007-c25e-42a2-a7f5-fea2131d154c" containerID="5dab16f693141f9da639930da2bd902921b2215db0bd5052ef801cf4ec8eec67" exitCode=0 Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.625312 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerDied","Data":"5dab16f693141f9da639930da2bd902921b2215db0bd5052ef801cf4ec8eec67"} Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.625358 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cr597" event={"ID":"01715007-c25e-42a2-a7f5-fea2131d154c","Type":"ContainerDied","Data":"8b458169847cbaf35aad300f418d9325a9f4c0b66c9c46a1b0338de235834a1b"} Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.625382 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b458169847cbaf35aad300f418d9325a9f4c0b66c9c46a1b0338de235834a1b" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.764375 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.797646 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-catalog-content\") pod \"01715007-c25e-42a2-a7f5-fea2131d154c\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.797725 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-utilities\") pod \"01715007-c25e-42a2-a7f5-fea2131d154c\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.797792 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67z44\" (UniqueName: \"kubernetes.io/projected/01715007-c25e-42a2-a7f5-fea2131d154c-kube-api-access-67z44\") pod \"01715007-c25e-42a2-a7f5-fea2131d154c\" (UID: \"01715007-c25e-42a2-a7f5-fea2131d154c\") " Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.798594 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-utilities" (OuterVolumeSpecName: "utilities") pod "01715007-c25e-42a2-a7f5-fea2131d154c" (UID: "01715007-c25e-42a2-a7f5-fea2131d154c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.813292 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01715007-c25e-42a2-a7f5-fea2131d154c-kube-api-access-67z44" (OuterVolumeSpecName: "kube-api-access-67z44") pod "01715007-c25e-42a2-a7f5-fea2131d154c" (UID: "01715007-c25e-42a2-a7f5-fea2131d154c"). InnerVolumeSpecName "kube-api-access-67z44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.889649 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "01715007-c25e-42a2-a7f5-fea2131d154c" (UID: "01715007-c25e-42a2-a7f5-fea2131d154c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.901086 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.901122 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01715007-c25e-42a2-a7f5-fea2131d154c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:38:39 crc kubenswrapper[4829]: I1002 08:38:39.901132 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67z44\" (UniqueName: \"kubernetes.io/projected/01715007-c25e-42a2-a7f5-fea2131d154c-kube-api-access-67z44\") on node \"crc\" DevicePath \"\"" Oct 02 08:38:40 crc kubenswrapper[4829]: I1002 08:38:40.635244 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cr597" Oct 02 08:38:40 crc kubenswrapper[4829]: I1002 08:38:40.671738 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cr597"] Oct 02 08:38:40 crc kubenswrapper[4829]: I1002 08:38:40.682533 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cr597"] Oct 02 08:38:41 crc kubenswrapper[4829]: I1002 08:38:41.484180 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" path="/var/lib/kubelet/pods/01715007-c25e-42a2-a7f5-fea2131d154c/volumes" Oct 02 08:38:55 crc kubenswrapper[4829]: I1002 08:38:55.329335 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:38:55 crc kubenswrapper[4829]: I1002 08:38:55.329972 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:39:25 crc kubenswrapper[4829]: I1002 08:39:25.329063 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:39:25 crc kubenswrapper[4829]: I1002 08:39:25.329724 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:39:25 crc kubenswrapper[4829]: I1002 08:39:25.329785 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:39:25 crc kubenswrapper[4829]: I1002 08:39:25.330896 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:39:25 crc kubenswrapper[4829]: I1002 08:39:25.331028 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" gracePeriod=600 Oct 02 08:39:25 crc kubenswrapper[4829]: E1002 08:39:25.475573 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:39:26 crc kubenswrapper[4829]: I1002 08:39:26.134378 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" exitCode=0 Oct 02 08:39:26 crc kubenswrapper[4829]: I1002 08:39:26.134464 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5"} Oct 02 08:39:26 crc kubenswrapper[4829]: I1002 08:39:26.134737 4829 scope.go:117] "RemoveContainer" containerID="d9daaaad0373f37594d146a03b1ae22f94640ac1896cf827746fd46a113423ab" Oct 02 08:39:26 crc kubenswrapper[4829]: I1002 08:39:26.135677 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:39:26 crc kubenswrapper[4829]: E1002 08:39:26.136247 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:39:36 crc kubenswrapper[4829]: I1002 08:39:36.461707 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:39:36 crc kubenswrapper[4829]: E1002 08:39:36.462859 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:39:49 crc kubenswrapper[4829]: I1002 08:39:49.468853 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:39:49 crc kubenswrapper[4829]: E1002 08:39:49.469725 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:40:04 crc kubenswrapper[4829]: I1002 08:40:04.460989 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:40:04 crc kubenswrapper[4829]: E1002 08:40:04.461840 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:40:17 crc kubenswrapper[4829]: I1002 08:40:17.461715 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:40:17 crc kubenswrapper[4829]: E1002 08:40:17.462435 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:40:28 crc kubenswrapper[4829]: I1002 08:40:28.461908 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:40:28 crc kubenswrapper[4829]: E1002 08:40:28.462920 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:40:41 crc kubenswrapper[4829]: I1002 08:40:41.462787 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:40:41 crc kubenswrapper[4829]: E1002 08:40:41.464220 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:40:52 crc kubenswrapper[4829]: I1002 08:40:52.461521 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:40:52 crc kubenswrapper[4829]: E1002 08:40:52.462435 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:41:07 crc kubenswrapper[4829]: I1002 08:41:07.461675 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:41:07 crc kubenswrapper[4829]: E1002 08:41:07.462547 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.851071 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q8fgd"] Oct 02 08:41:19 crc kubenswrapper[4829]: E1002 08:41:19.851933 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="extract-utilities" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.851945 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="extract-utilities" Oct 02 08:41:19 crc kubenswrapper[4829]: E1002 08:41:19.851985 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="registry-server" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.851993 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="registry-server" Oct 02 08:41:19 crc kubenswrapper[4829]: E1002 08:41:19.852008 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="extract-content" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.852014 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="extract-content" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.852194 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="01715007-c25e-42a2-a7f5-fea2131d154c" containerName="registry-server" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.853671 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.877248 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q8fgd"] Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.945107 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-utilities\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.945458 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqtg4\" (UniqueName: \"kubernetes.io/projected/0b409ab0-690f-41f8-981e-9d3f3b123e90-kube-api-access-mqtg4\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:19 crc kubenswrapper[4829]: I1002 08:41:19.945662 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-catalog-content\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.051704 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-catalog-content\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.052172 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-catalog-content\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.052684 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-utilities\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.052969 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-utilities\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.053192 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqtg4\" (UniqueName: \"kubernetes.io/projected/0b409ab0-690f-41f8-981e-9d3f3b123e90-kube-api-access-mqtg4\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.085730 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqtg4\" (UniqueName: \"kubernetes.io/projected/0b409ab0-690f-41f8-981e-9d3f3b123e90-kube-api-access-mqtg4\") pod \"certified-operators-q8fgd\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.172884 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:20 crc kubenswrapper[4829]: I1002 08:41:20.756254 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q8fgd"] Oct 02 08:41:21 crc kubenswrapper[4829]: I1002 08:41:21.442070 4829 generic.go:334] "Generic (PLEG): container finished" podID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerID="624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e" exitCode=0 Oct 02 08:41:21 crc kubenswrapper[4829]: I1002 08:41:21.442157 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerDied","Data":"624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e"} Oct 02 08:41:21 crc kubenswrapper[4829]: I1002 08:41:21.442206 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerStarted","Data":"d35293317bd2fae29979db994eb70d6aabbab52fea5dcf5def949a025aecb79e"} Oct 02 08:41:21 crc kubenswrapper[4829]: I1002 08:41:21.462258 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:41:21 crc kubenswrapper[4829]: E1002 08:41:21.463609 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:41:23 crc kubenswrapper[4829]: I1002 08:41:23.480569 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerStarted","Data":"32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce"} Oct 02 08:41:24 crc kubenswrapper[4829]: I1002 08:41:24.491866 4829 generic.go:334] "Generic (PLEG): container finished" podID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerID="32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce" exitCode=0 Oct 02 08:41:24 crc kubenswrapper[4829]: I1002 08:41:24.491991 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerDied","Data":"32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce"} Oct 02 08:41:25 crc kubenswrapper[4829]: I1002 08:41:25.503556 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerStarted","Data":"5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf"} Oct 02 08:41:25 crc kubenswrapper[4829]: I1002 08:41:25.525193 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q8fgd" podStartSLOduration=3.058410628 podStartE2EDuration="6.525174949s" podCreationTimestamp="2025-10-02 08:41:19 +0000 UTC" firstStartedPulling="2025-10-02 08:41:21.445208718 +0000 UTC m=+5072.784857173" lastFinishedPulling="2025-10-02 08:41:24.911973049 +0000 UTC m=+5076.251621494" observedRunningTime="2025-10-02 08:41:25.523933612 +0000 UTC m=+5076.863582017" watchObservedRunningTime="2025-10-02 08:41:25.525174949 +0000 UTC m=+5076.864823364" Oct 02 08:41:30 crc kubenswrapper[4829]: I1002 08:41:30.173451 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:30 crc kubenswrapper[4829]: I1002 08:41:30.173892 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:30 crc kubenswrapper[4829]: I1002 08:41:30.264942 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:30 crc kubenswrapper[4829]: I1002 08:41:30.646519 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:30 crc kubenswrapper[4829]: I1002 08:41:30.714601 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q8fgd"] Oct 02 08:41:32 crc kubenswrapper[4829]: I1002 08:41:32.584009 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q8fgd" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="registry-server" containerID="cri-o://5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf" gracePeriod=2 Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.131158 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.275161 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-utilities\") pod \"0b409ab0-690f-41f8-981e-9d3f3b123e90\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.275333 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-catalog-content\") pod \"0b409ab0-690f-41f8-981e-9d3f3b123e90\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.275486 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqtg4\" (UniqueName: \"kubernetes.io/projected/0b409ab0-690f-41f8-981e-9d3f3b123e90-kube-api-access-mqtg4\") pod \"0b409ab0-690f-41f8-981e-9d3f3b123e90\" (UID: \"0b409ab0-690f-41f8-981e-9d3f3b123e90\") " Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.276029 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-utilities" (OuterVolumeSpecName: "utilities") pod "0b409ab0-690f-41f8-981e-9d3f3b123e90" (UID: "0b409ab0-690f-41f8-981e-9d3f3b123e90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.286088 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b409ab0-690f-41f8-981e-9d3f3b123e90-kube-api-access-mqtg4" (OuterVolumeSpecName: "kube-api-access-mqtg4") pod "0b409ab0-690f-41f8-981e-9d3f3b123e90" (UID: "0b409ab0-690f-41f8-981e-9d3f3b123e90"). InnerVolumeSpecName "kube-api-access-mqtg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.318810 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b409ab0-690f-41f8-981e-9d3f3b123e90" (UID: "0b409ab0-690f-41f8-981e-9d3f3b123e90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.377825 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.377873 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqtg4\" (UniqueName: \"kubernetes.io/projected/0b409ab0-690f-41f8-981e-9d3f3b123e90-kube-api-access-mqtg4\") on node \"crc\" DevicePath \"\"" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.377892 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b409ab0-690f-41f8-981e-9d3f3b123e90-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.601394 4829 generic.go:334] "Generic (PLEG): container finished" podID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerID="5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf" exitCode=0 Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.601475 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerDied","Data":"5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf"} Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.601528 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q8fgd" event={"ID":"0b409ab0-690f-41f8-981e-9d3f3b123e90","Type":"ContainerDied","Data":"d35293317bd2fae29979db994eb70d6aabbab52fea5dcf5def949a025aecb79e"} Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.601564 4829 scope.go:117] "RemoveContainer" containerID="5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.601799 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q8fgd" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.638834 4829 scope.go:117] "RemoveContainer" containerID="32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.650141 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q8fgd"] Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.657790 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q8fgd"] Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.664607 4829 scope.go:117] "RemoveContainer" containerID="624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.728084 4829 scope.go:117] "RemoveContainer" containerID="5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf" Oct 02 08:41:33 crc kubenswrapper[4829]: E1002 08:41:33.728600 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf\": container with ID starting with 5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf not found: ID does not exist" containerID="5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.728639 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf"} err="failed to get container status \"5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf\": rpc error: code = NotFound desc = could not find container \"5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf\": container with ID starting with 5128287faeb1bcfdf6d079882156b745b1399c8a329a6302333f2584fb20f8cf not found: ID does not exist" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.728664 4829 scope.go:117] "RemoveContainer" containerID="32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce" Oct 02 08:41:33 crc kubenswrapper[4829]: E1002 08:41:33.729294 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce\": container with ID starting with 32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce not found: ID does not exist" containerID="32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.729337 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce"} err="failed to get container status \"32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce\": rpc error: code = NotFound desc = could not find container \"32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce\": container with ID starting with 32ce67aeb909ce9ba241e67cdd6e0b482a88472bbc1da1acab659f8a4c4e13ce not found: ID does not exist" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.729365 4829 scope.go:117] "RemoveContainer" containerID="624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e" Oct 02 08:41:33 crc kubenswrapper[4829]: E1002 08:41:33.729749 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e\": container with ID starting with 624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e not found: ID does not exist" containerID="624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e" Oct 02 08:41:33 crc kubenswrapper[4829]: I1002 08:41:33.729775 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e"} err="failed to get container status \"624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e\": rpc error: code = NotFound desc = could not find container \"624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e\": container with ID starting with 624c221d6e3eb9e5e962ab360a5d67caf495b89df13790b1ec782d7fff02d54e not found: ID does not exist" Oct 02 08:41:34 crc kubenswrapper[4829]: I1002 08:41:34.461126 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:41:34 crc kubenswrapper[4829]: E1002 08:41:34.462306 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:41:35 crc kubenswrapper[4829]: I1002 08:41:35.477136 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" path="/var/lib/kubelet/pods/0b409ab0-690f-41f8-981e-9d3f3b123e90/volumes" Oct 02 08:41:47 crc kubenswrapper[4829]: I1002 08:41:47.461562 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:41:47 crc kubenswrapper[4829]: E1002 08:41:47.462615 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:42:01 crc kubenswrapper[4829]: I1002 08:42:01.461268 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:42:01 crc kubenswrapper[4829]: E1002 08:42:01.462446 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:42:12 crc kubenswrapper[4829]: I1002 08:42:12.461768 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:42:12 crc kubenswrapper[4829]: E1002 08:42:12.462965 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.447452 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rq94k"] Oct 02 08:42:22 crc kubenswrapper[4829]: E1002 08:42:22.449833 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="registry-server" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.449874 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="registry-server" Oct 02 08:42:22 crc kubenswrapper[4829]: E1002 08:42:22.449949 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="extract-utilities" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.449971 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="extract-utilities" Oct 02 08:42:22 crc kubenswrapper[4829]: E1002 08:42:22.450043 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="extract-content" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.450058 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="extract-content" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.450907 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b409ab0-690f-41f8-981e-9d3f3b123e90" containerName="registry-server" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.458478 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.491050 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rq94k"] Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.528169 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-utilities\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.528785 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-catalog-content\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.529542 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dxd\" (UniqueName: \"kubernetes.io/projected/33593175-129d-448b-835a-aec5024ee8aa-kube-api-access-g8dxd\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.632592 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-utilities\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.632901 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-catalog-content\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.632987 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dxd\" (UniqueName: \"kubernetes.io/projected/33593175-129d-448b-835a-aec5024ee8aa-kube-api-access-g8dxd\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.633143 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-utilities\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.633483 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-catalog-content\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.663817 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dxd\" (UniqueName: \"kubernetes.io/projected/33593175-129d-448b-835a-aec5024ee8aa-kube-api-access-g8dxd\") pod \"redhat-operators-rq94k\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:22 crc kubenswrapper[4829]: I1002 08:42:22.813816 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:23 crc kubenswrapper[4829]: I1002 08:42:23.339936 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rq94k"] Oct 02 08:42:24 crc kubenswrapper[4829]: I1002 08:42:24.157780 4829 generic.go:334] "Generic (PLEG): container finished" podID="33593175-129d-448b-835a-aec5024ee8aa" containerID="91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656" exitCode=0 Oct 02 08:42:24 crc kubenswrapper[4829]: I1002 08:42:24.157862 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerDied","Data":"91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656"} Oct 02 08:42:24 crc kubenswrapper[4829]: I1002 08:42:24.158172 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerStarted","Data":"c14b99deb61e4e404190b707203dabf70fe6febd87a144210124be33a3c95eb4"} Oct 02 08:42:24 crc kubenswrapper[4829]: I1002 08:42:24.461985 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:42:24 crc kubenswrapper[4829]: E1002 08:42:24.462294 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:42:28 crc kubenswrapper[4829]: I1002 08:42:28.200064 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerStarted","Data":"ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73"} Oct 02 08:42:39 crc kubenswrapper[4829]: I1002 08:42:39.473609 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:42:39 crc kubenswrapper[4829]: E1002 08:42:39.474809 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:42:41 crc kubenswrapper[4829]: I1002 08:42:41.340666 4829 generic.go:334] "Generic (PLEG): container finished" podID="33593175-129d-448b-835a-aec5024ee8aa" containerID="ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73" exitCode=0 Oct 02 08:42:41 crc kubenswrapper[4829]: I1002 08:42:41.340775 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerDied","Data":"ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73"} Oct 02 08:42:43 crc kubenswrapper[4829]: I1002 08:42:43.368389 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerStarted","Data":"273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d"} Oct 02 08:42:43 crc kubenswrapper[4829]: I1002 08:42:43.390447 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rq94k" podStartSLOduration=3.023894528 podStartE2EDuration="21.390425172s" podCreationTimestamp="2025-10-02 08:42:22 +0000 UTC" firstStartedPulling="2025-10-02 08:42:24.16110309 +0000 UTC m=+5135.500751505" lastFinishedPulling="2025-10-02 08:42:42.527633734 +0000 UTC m=+5153.867282149" observedRunningTime="2025-10-02 08:42:43.390143243 +0000 UTC m=+5154.729791658" watchObservedRunningTime="2025-10-02 08:42:43.390425172 +0000 UTC m=+5154.730073587" Oct 02 08:42:50 crc kubenswrapper[4829]: I1002 08:42:50.461351 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:42:50 crc kubenswrapper[4829]: E1002 08:42:50.462282 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:42:52 crc kubenswrapper[4829]: I1002 08:42:52.813973 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:52 crc kubenswrapper[4829]: I1002 08:42:52.815547 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:52 crc kubenswrapper[4829]: I1002 08:42:52.894840 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:53 crc kubenswrapper[4829]: I1002 08:42:53.540731 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:53 crc kubenswrapper[4829]: I1002 08:42:53.641037 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rq94k"] Oct 02 08:42:55 crc kubenswrapper[4829]: I1002 08:42:55.497393 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rq94k" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="registry-server" containerID="cri-o://273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d" gracePeriod=2 Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.065522 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.100906 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-utilities\") pod \"33593175-129d-448b-835a-aec5024ee8aa\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.101218 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8dxd\" (UniqueName: \"kubernetes.io/projected/33593175-129d-448b-835a-aec5024ee8aa-kube-api-access-g8dxd\") pod \"33593175-129d-448b-835a-aec5024ee8aa\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.101505 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-catalog-content\") pod \"33593175-129d-448b-835a-aec5024ee8aa\" (UID: \"33593175-129d-448b-835a-aec5024ee8aa\") " Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.102755 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-utilities" (OuterVolumeSpecName: "utilities") pod "33593175-129d-448b-835a-aec5024ee8aa" (UID: "33593175-129d-448b-835a-aec5024ee8aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.107380 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33593175-129d-448b-835a-aec5024ee8aa-kube-api-access-g8dxd" (OuterVolumeSpecName: "kube-api-access-g8dxd") pod "33593175-129d-448b-835a-aec5024ee8aa" (UID: "33593175-129d-448b-835a-aec5024ee8aa"). InnerVolumeSpecName "kube-api-access-g8dxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.204296 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.204335 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8dxd\" (UniqueName: \"kubernetes.io/projected/33593175-129d-448b-835a-aec5024ee8aa-kube-api-access-g8dxd\") on node \"crc\" DevicePath \"\"" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.205110 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "33593175-129d-448b-835a-aec5024ee8aa" (UID: "33593175-129d-448b-835a-aec5024ee8aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.306008 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/33593175-129d-448b-835a-aec5024ee8aa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.510447 4829 generic.go:334] "Generic (PLEG): container finished" podID="33593175-129d-448b-835a-aec5024ee8aa" containerID="273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d" exitCode=0 Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.510502 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerDied","Data":"273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d"} Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.510558 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rq94k" event={"ID":"33593175-129d-448b-835a-aec5024ee8aa","Type":"ContainerDied","Data":"c14b99deb61e4e404190b707203dabf70fe6febd87a144210124be33a3c95eb4"} Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.510573 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rq94k" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.510585 4829 scope.go:117] "RemoveContainer" containerID="273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.562081 4829 scope.go:117] "RemoveContainer" containerID="ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.566465 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rq94k"] Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.579122 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rq94k"] Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.915613 4829 scope.go:117] "RemoveContainer" containerID="91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.984570 4829 scope.go:117] "RemoveContainer" containerID="273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d" Oct 02 08:42:56 crc kubenswrapper[4829]: E1002 08:42:56.985265 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d\": container with ID starting with 273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d not found: ID does not exist" containerID="273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.985328 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d"} err="failed to get container status \"273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d\": rpc error: code = NotFound desc = could not find container \"273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d\": container with ID starting with 273c968e71c9ab75918ee40cf2bdb2e084c94305cdcb2b2fbcac614d8bf1be3d not found: ID does not exist" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.985367 4829 scope.go:117] "RemoveContainer" containerID="ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73" Oct 02 08:42:56 crc kubenswrapper[4829]: E1002 08:42:56.985829 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73\": container with ID starting with ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73 not found: ID does not exist" containerID="ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.985869 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73"} err="failed to get container status \"ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73\": rpc error: code = NotFound desc = could not find container \"ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73\": container with ID starting with ab98db8fb60ca08b05c0bea67ea9d03e1e1823f25fae0bddb35cad9a4c432e73 not found: ID does not exist" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.985895 4829 scope.go:117] "RemoveContainer" containerID="91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656" Oct 02 08:42:56 crc kubenswrapper[4829]: E1002 08:42:56.986365 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656\": container with ID starting with 91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656 not found: ID does not exist" containerID="91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656" Oct 02 08:42:56 crc kubenswrapper[4829]: I1002 08:42:56.986396 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656"} err="failed to get container status \"91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656\": rpc error: code = NotFound desc = could not find container \"91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656\": container with ID starting with 91bb254be0731d86a9d79884f7bbd10ee0daeef7ff9f077b537e7ca3df0b9656 not found: ID does not exist" Oct 02 08:42:57 crc kubenswrapper[4829]: I1002 08:42:57.483460 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33593175-129d-448b-835a-aec5024ee8aa" path="/var/lib/kubelet/pods/33593175-129d-448b-835a-aec5024ee8aa/volumes" Oct 02 08:43:05 crc kubenswrapper[4829]: I1002 08:43:05.461562 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:43:05 crc kubenswrapper[4829]: E1002 08:43:05.462356 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:43:18 crc kubenswrapper[4829]: I1002 08:43:18.461500 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:43:18 crc kubenswrapper[4829]: E1002 08:43:18.462243 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:43:29 crc kubenswrapper[4829]: I1002 08:43:29.472553 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:43:29 crc kubenswrapper[4829]: E1002 08:43:29.474114 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:43:41 crc kubenswrapper[4829]: I1002 08:43:41.461288 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:43:41 crc kubenswrapper[4829]: E1002 08:43:41.462267 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:43:53 crc kubenswrapper[4829]: I1002 08:43:53.461987 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:43:53 crc kubenswrapper[4829]: E1002 08:43:53.462763 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:44:04 crc kubenswrapper[4829]: I1002 08:44:04.461456 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:44:04 crc kubenswrapper[4829]: E1002 08:44:04.462638 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:44:19 crc kubenswrapper[4829]: I1002 08:44:19.466453 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:44:19 crc kubenswrapper[4829]: E1002 08:44:19.467143 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:44:33 crc kubenswrapper[4829]: I1002 08:44:33.461615 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:44:33 crc kubenswrapper[4829]: I1002 08:44:33.750649 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"fd55a71d4738705ab0f70625de7b8981cb5b16abfb30540fbfe73c64ac1d1a62"} Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.156756 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k"] Oct 02 08:45:00 crc kubenswrapper[4829]: E1002 08:45:00.157937 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="extract-utilities" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.157956 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="extract-utilities" Oct 02 08:45:00 crc kubenswrapper[4829]: E1002 08:45:00.157976 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="registry-server" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.157986 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="registry-server" Oct 02 08:45:00 crc kubenswrapper[4829]: E1002 08:45:00.158009 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="extract-content" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.158019 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="extract-content" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.158327 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="33593175-129d-448b-835a-aec5024ee8aa" containerName="registry-server" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.159101 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.160664 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.161927 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.166709 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k"] Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.245817 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/958c5c65-87b7-4665-a4b7-7e34d7e7224d-config-volume\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.246188 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw26x\" (UniqueName: \"kubernetes.io/projected/958c5c65-87b7-4665-a4b7-7e34d7e7224d-kube-api-access-bw26x\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.246309 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/958c5c65-87b7-4665-a4b7-7e34d7e7224d-secret-volume\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.347888 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/958c5c65-87b7-4665-a4b7-7e34d7e7224d-config-volume\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.348048 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw26x\" (UniqueName: \"kubernetes.io/projected/958c5c65-87b7-4665-a4b7-7e34d7e7224d-kube-api-access-bw26x\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.348077 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/958c5c65-87b7-4665-a4b7-7e34d7e7224d-secret-volume\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.348862 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/958c5c65-87b7-4665-a4b7-7e34d7e7224d-config-volume\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.353894 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/958c5c65-87b7-4665-a4b7-7e34d7e7224d-secret-volume\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.370284 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw26x\" (UniqueName: \"kubernetes.io/projected/958c5c65-87b7-4665-a4b7-7e34d7e7224d-kube-api-access-bw26x\") pod \"collect-profiles-29323245-24s8k\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.485839 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.838635 4829 scope.go:117] "RemoveContainer" containerID="5dab16f693141f9da639930da2bd902921b2215db0bd5052ef801cf4ec8eec67" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.872385 4829 scope.go:117] "RemoveContainer" containerID="0b5a54e38e8450a389d4f16b45b2e9c05d10cd3f2917f1a6f2e678ceab48253e" Oct 02 08:45:00 crc kubenswrapper[4829]: I1002 08:45:00.912495 4829 scope.go:117] "RemoveContainer" containerID="27fc2a7d7934d2615b7815106b6b67fabd606269d639ff709656e39e47d802c9" Oct 02 08:45:01 crc kubenswrapper[4829]: I1002 08:45:01.028101 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k"] Oct 02 08:45:01 crc kubenswrapper[4829]: I1002 08:45:01.105412 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" event={"ID":"958c5c65-87b7-4665-a4b7-7e34d7e7224d","Type":"ContainerStarted","Data":"e6c2bae99f5e8f2b73852b37c2b4d8a7c3f473535e94b84ffa8841d3aad7472c"} Oct 02 08:45:02 crc kubenswrapper[4829]: I1002 08:45:02.117049 4829 generic.go:334] "Generic (PLEG): container finished" podID="958c5c65-87b7-4665-a4b7-7e34d7e7224d" containerID="4e813c3feaa02975952b0c8fb21f62a9d3b0e2a0e7f707f8fec7c43694252a39" exitCode=0 Oct 02 08:45:02 crc kubenswrapper[4829]: I1002 08:45:02.117126 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" event={"ID":"958c5c65-87b7-4665-a4b7-7e34d7e7224d","Type":"ContainerDied","Data":"4e813c3feaa02975952b0c8fb21f62a9d3b0e2a0e7f707f8fec7c43694252a39"} Oct 02 08:45:03 crc kubenswrapper[4829]: I1002 08:45:03.521794 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:03 crc kubenswrapper[4829]: I1002 08:45:03.613512 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/958c5c65-87b7-4665-a4b7-7e34d7e7224d-secret-volume\") pod \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " Oct 02 08:45:03 crc kubenswrapper[4829]: I1002 08:45:03.613571 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/958c5c65-87b7-4665-a4b7-7e34d7e7224d-config-volume\") pod \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " Oct 02 08:45:03 crc kubenswrapper[4829]: I1002 08:45:03.613620 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw26x\" (UniqueName: \"kubernetes.io/projected/958c5c65-87b7-4665-a4b7-7e34d7e7224d-kube-api-access-bw26x\") pod \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\" (UID: \"958c5c65-87b7-4665-a4b7-7e34d7e7224d\") " Oct 02 08:45:03 crc kubenswrapper[4829]: I1002 08:45:03.615116 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/958c5c65-87b7-4665-a4b7-7e34d7e7224d-config-volume" (OuterVolumeSpecName: "config-volume") pod "958c5c65-87b7-4665-a4b7-7e34d7e7224d" (UID: "958c5c65-87b7-4665-a4b7-7e34d7e7224d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:45:03 crc kubenswrapper[4829]: I1002 08:45:03.716015 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/958c5c65-87b7-4665-a4b7-7e34d7e7224d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.142685 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" event={"ID":"958c5c65-87b7-4665-a4b7-7e34d7e7224d","Type":"ContainerDied","Data":"e6c2bae99f5e8f2b73852b37c2b4d8a7c3f473535e94b84ffa8841d3aad7472c"} Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.142968 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6c2bae99f5e8f2b73852b37c2b4d8a7c3f473535e94b84ffa8841d3aad7472c" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.142758 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323245-24s8k" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.304817 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/958c5c65-87b7-4665-a4b7-7e34d7e7224d-kube-api-access-bw26x" (OuterVolumeSpecName: "kube-api-access-bw26x") pod "958c5c65-87b7-4665-a4b7-7e34d7e7224d" (UID: "958c5c65-87b7-4665-a4b7-7e34d7e7224d"). InnerVolumeSpecName "kube-api-access-bw26x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.304995 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/958c5c65-87b7-4665-a4b7-7e34d7e7224d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "958c5c65-87b7-4665-a4b7-7e34d7e7224d" (UID: "958c5c65-87b7-4665-a4b7-7e34d7e7224d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.332084 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/958c5c65-87b7-4665-a4b7-7e34d7e7224d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.332410 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw26x\" (UniqueName: \"kubernetes.io/projected/958c5c65-87b7-4665-a4b7-7e34d7e7224d-kube-api-access-bw26x\") on node \"crc\" DevicePath \"\"" Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.598506 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl"] Oct 02 08:45:04 crc kubenswrapper[4829]: I1002 08:45:04.605670 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323200-nqwsl"] Oct 02 08:45:05 crc kubenswrapper[4829]: I1002 08:45:05.475330 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e33017a-e2ee-470c-b9e5-03b4602ff7a0" path="/var/lib/kubelet/pods/4e33017a-e2ee-470c-b9e5-03b4602ff7a0/volumes" Oct 02 08:46:01 crc kubenswrapper[4829]: I1002 08:46:01.021798 4829 scope.go:117] "RemoveContainer" containerID="a705f53b9287a8c09aedb390e28d761082a3ca0779d24d5fc09c4fe742410d32" Oct 02 08:46:55 crc kubenswrapper[4829]: I1002 08:46:55.329349 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:46:55 crc kubenswrapper[4829]: I1002 08:46:55.330297 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:47:23 crc kubenswrapper[4829]: I1002 08:47:23.958901 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jqr86"] Oct 02 08:47:23 crc kubenswrapper[4829]: E1002 08:47:23.962788 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="958c5c65-87b7-4665-a4b7-7e34d7e7224d" containerName="collect-profiles" Oct 02 08:47:23 crc kubenswrapper[4829]: I1002 08:47:23.962973 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="958c5c65-87b7-4665-a4b7-7e34d7e7224d" containerName="collect-profiles" Oct 02 08:47:23 crc kubenswrapper[4829]: I1002 08:47:23.963828 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="958c5c65-87b7-4665-a4b7-7e34d7e7224d" containerName="collect-profiles" Oct 02 08:47:23 crc kubenswrapper[4829]: I1002 08:47:23.969429 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:23 crc kubenswrapper[4829]: I1002 08:47:23.994668 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqr86"] Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.154023 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-catalog-content\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.154097 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6k5l\" (UniqueName: \"kubernetes.io/projected/44a874d9-38bf-458e-b986-6225e540421d-kube-api-access-n6k5l\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.154157 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-utilities\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.255802 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-utilities\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.255967 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-catalog-content\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.256002 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6k5l\" (UniqueName: \"kubernetes.io/projected/44a874d9-38bf-458e-b986-6225e540421d-kube-api-access-n6k5l\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.256752 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-utilities\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.256912 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-catalog-content\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.282481 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6k5l\" (UniqueName: \"kubernetes.io/projected/44a874d9-38bf-458e-b986-6225e540421d-kube-api-access-n6k5l\") pod \"redhat-marketplace-jqr86\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.316803 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.792524 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqr86"] Oct 02 08:47:24 crc kubenswrapper[4829]: I1002 08:47:24.816900 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerStarted","Data":"de16f9eb5364d81a2523e5c16326dff5a3cf179e1d46aac035d6a76918feb591"} Oct 02 08:47:25 crc kubenswrapper[4829]: I1002 08:47:25.329897 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:47:25 crc kubenswrapper[4829]: I1002 08:47:25.330517 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:47:25 crc kubenswrapper[4829]: I1002 08:47:25.830010 4829 generic.go:334] "Generic (PLEG): container finished" podID="44a874d9-38bf-458e-b986-6225e540421d" containerID="d1db2b893a28c7d894fc4fd7db90b2bd511700b6eba9b24fa8ca6def0e4a3321" exitCode=0 Oct 02 08:47:25 crc kubenswrapper[4829]: I1002 08:47:25.830084 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerDied","Data":"d1db2b893a28c7d894fc4fd7db90b2bd511700b6eba9b24fa8ca6def0e4a3321"} Oct 02 08:47:25 crc kubenswrapper[4829]: I1002 08:47:25.834701 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:47:26 crc kubenswrapper[4829]: I1002 08:47:26.846878 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerStarted","Data":"38543c24db83a4b3ee983dce25838c71cf0fc649c8a98aff0b14bbdef8feaa87"} Oct 02 08:47:27 crc kubenswrapper[4829]: I1002 08:47:27.872093 4829 generic.go:334] "Generic (PLEG): container finished" podID="44a874d9-38bf-458e-b986-6225e540421d" containerID="38543c24db83a4b3ee983dce25838c71cf0fc649c8a98aff0b14bbdef8feaa87" exitCode=0 Oct 02 08:47:27 crc kubenswrapper[4829]: I1002 08:47:27.872153 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerDied","Data":"38543c24db83a4b3ee983dce25838c71cf0fc649c8a98aff0b14bbdef8feaa87"} Oct 02 08:47:28 crc kubenswrapper[4829]: I1002 08:47:28.882464 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerStarted","Data":"2c1d7f541aee8a8a646115baa4f41309b5d3426c46534d44d6105e18157c331e"} Oct 02 08:47:28 crc kubenswrapper[4829]: I1002 08:47:28.904287 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jqr86" podStartSLOduration=3.306591462 podStartE2EDuration="5.904268056s" podCreationTimestamp="2025-10-02 08:47:23 +0000 UTC" firstStartedPulling="2025-10-02 08:47:25.834278365 +0000 UTC m=+5437.173926810" lastFinishedPulling="2025-10-02 08:47:28.431954989 +0000 UTC m=+5439.771603404" observedRunningTime="2025-10-02 08:47:28.898272835 +0000 UTC m=+5440.237921250" watchObservedRunningTime="2025-10-02 08:47:28.904268056 +0000 UTC m=+5440.243916461" Oct 02 08:47:34 crc kubenswrapper[4829]: I1002 08:47:34.317781 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:34 crc kubenswrapper[4829]: I1002 08:47:34.318438 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:34 crc kubenswrapper[4829]: I1002 08:47:34.379542 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:35 crc kubenswrapper[4829]: I1002 08:47:35.566588 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:35 crc kubenswrapper[4829]: I1002 08:47:35.639189 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqr86"] Oct 02 08:47:36 crc kubenswrapper[4829]: I1002 08:47:36.995473 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jqr86" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="registry-server" containerID="cri-o://2c1d7f541aee8a8a646115baa4f41309b5d3426c46534d44d6105e18157c331e" gracePeriod=2 Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.007909 4829 generic.go:334] "Generic (PLEG): container finished" podID="44a874d9-38bf-458e-b986-6225e540421d" containerID="2c1d7f541aee8a8a646115baa4f41309b5d3426c46534d44d6105e18157c331e" exitCode=0 Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.008029 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerDied","Data":"2c1d7f541aee8a8a646115baa4f41309b5d3426c46534d44d6105e18157c331e"} Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.008319 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jqr86" event={"ID":"44a874d9-38bf-458e-b986-6225e540421d","Type":"ContainerDied","Data":"de16f9eb5364d81a2523e5c16326dff5a3cf179e1d46aac035d6a76918feb591"} Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.008347 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de16f9eb5364d81a2523e5c16326dff5a3cf179e1d46aac035d6a76918feb591" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.038691 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.188298 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-utilities\") pod \"44a874d9-38bf-458e-b986-6225e540421d\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.188859 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-catalog-content\") pod \"44a874d9-38bf-458e-b986-6225e540421d\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.188921 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6k5l\" (UniqueName: \"kubernetes.io/projected/44a874d9-38bf-458e-b986-6225e540421d-kube-api-access-n6k5l\") pod \"44a874d9-38bf-458e-b986-6225e540421d\" (UID: \"44a874d9-38bf-458e-b986-6225e540421d\") " Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.189207 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-utilities" (OuterVolumeSpecName: "utilities") pod "44a874d9-38bf-458e-b986-6225e540421d" (UID: "44a874d9-38bf-458e-b986-6225e540421d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.189949 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.198192 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44a874d9-38bf-458e-b986-6225e540421d-kube-api-access-n6k5l" (OuterVolumeSpecName: "kube-api-access-n6k5l") pod "44a874d9-38bf-458e-b986-6225e540421d" (UID: "44a874d9-38bf-458e-b986-6225e540421d"). InnerVolumeSpecName "kube-api-access-n6k5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.203432 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44a874d9-38bf-458e-b986-6225e540421d" (UID: "44a874d9-38bf-458e-b986-6225e540421d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.291854 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44a874d9-38bf-458e-b986-6225e540421d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:38 crc kubenswrapper[4829]: I1002 08:47:38.291899 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6k5l\" (UniqueName: \"kubernetes.io/projected/44a874d9-38bf-458e-b986-6225e540421d-kube-api-access-n6k5l\") on node \"crc\" DevicePath \"\"" Oct 02 08:47:39 crc kubenswrapper[4829]: I1002 08:47:39.021469 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jqr86" Oct 02 08:47:39 crc kubenswrapper[4829]: I1002 08:47:39.071612 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqr86"] Oct 02 08:47:39 crc kubenswrapper[4829]: I1002 08:47:39.087030 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jqr86"] Oct 02 08:47:39 crc kubenswrapper[4829]: I1002 08:47:39.492409 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44a874d9-38bf-458e-b986-6225e540421d" path="/var/lib/kubelet/pods/44a874d9-38bf-458e-b986-6225e540421d/volumes" Oct 02 08:47:55 crc kubenswrapper[4829]: I1002 08:47:55.329635 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:47:55 crc kubenswrapper[4829]: I1002 08:47:55.330148 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:47:55 crc kubenswrapper[4829]: I1002 08:47:55.330203 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:47:55 crc kubenswrapper[4829]: I1002 08:47:55.331000 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fd55a71d4738705ab0f70625de7b8981cb5b16abfb30540fbfe73c64ac1d1a62"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:47:55 crc kubenswrapper[4829]: I1002 08:47:55.331053 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://fd55a71d4738705ab0f70625de7b8981cb5b16abfb30540fbfe73c64ac1d1a62" gracePeriod=600 Oct 02 08:47:56 crc kubenswrapper[4829]: I1002 08:47:56.234998 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="fd55a71d4738705ab0f70625de7b8981cb5b16abfb30540fbfe73c64ac1d1a62" exitCode=0 Oct 02 08:47:56 crc kubenswrapper[4829]: I1002 08:47:56.235062 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"fd55a71d4738705ab0f70625de7b8981cb5b16abfb30540fbfe73c64ac1d1a62"} Oct 02 08:47:56 crc kubenswrapper[4829]: I1002 08:47:56.235484 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592"} Oct 02 08:47:56 crc kubenswrapper[4829]: I1002 08:47:56.235504 4829 scope.go:117] "RemoveContainer" containerID="07d941dac054b1a5539f0010a8a001eb6ac0df6c588373ce743b21c297a1f9d5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.554642 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9mgs5"] Oct 02 08:48:40 crc kubenswrapper[4829]: E1002 08:48:40.555609 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="registry-server" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.555624 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="registry-server" Oct 02 08:48:40 crc kubenswrapper[4829]: E1002 08:48:40.555660 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="extract-utilities" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.555667 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="extract-utilities" Oct 02 08:48:40 crc kubenswrapper[4829]: E1002 08:48:40.555683 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="extract-content" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.555689 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="extract-content" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.555906 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="44a874d9-38bf-458e-b986-6225e540421d" containerName="registry-server" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.558139 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.583158 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9mgs5"] Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.702289 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t274q\" (UniqueName: \"kubernetes.io/projected/df4673d7-4d78-4e00-a610-38afd1c02f87-kube-api-access-t274q\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.702410 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-utilities\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.702458 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-catalog-content\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.804439 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-utilities\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.804766 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-catalog-content\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.804842 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t274q\" (UniqueName: \"kubernetes.io/projected/df4673d7-4d78-4e00-a610-38afd1c02f87-kube-api-access-t274q\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.805857 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-utilities\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.806125 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-catalog-content\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.824938 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t274q\" (UniqueName: \"kubernetes.io/projected/df4673d7-4d78-4e00-a610-38afd1c02f87-kube-api-access-t274q\") pod \"community-operators-9mgs5\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:40 crc kubenswrapper[4829]: I1002 08:48:40.889099 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:41 crc kubenswrapper[4829]: I1002 08:48:41.429690 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9mgs5"] Oct 02 08:48:41 crc kubenswrapper[4829]: W1002 08:48:41.431600 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf4673d7_4d78_4e00_a610_38afd1c02f87.slice/crio-3f4f8129dc949cdbd7ffaffb62e574fbc91812201564ea261c3a299b0c480cdf WatchSource:0}: Error finding container 3f4f8129dc949cdbd7ffaffb62e574fbc91812201564ea261c3a299b0c480cdf: Status 404 returned error can't find the container with id 3f4f8129dc949cdbd7ffaffb62e574fbc91812201564ea261c3a299b0c480cdf Oct 02 08:48:41 crc kubenswrapper[4829]: I1002 08:48:41.773293 4829 generic.go:334] "Generic (PLEG): container finished" podID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerID="45b27a8b9304a5e027b81d5c36a316b6dcc0e0dafcf014fd1f3ac5f5861d8f78" exitCode=0 Oct 02 08:48:41 crc kubenswrapper[4829]: I1002 08:48:41.773349 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerDied","Data":"45b27a8b9304a5e027b81d5c36a316b6dcc0e0dafcf014fd1f3ac5f5861d8f78"} Oct 02 08:48:41 crc kubenswrapper[4829]: I1002 08:48:41.773655 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerStarted","Data":"3f4f8129dc949cdbd7ffaffb62e574fbc91812201564ea261c3a299b0c480cdf"} Oct 02 08:48:43 crc kubenswrapper[4829]: I1002 08:48:43.801529 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerStarted","Data":"d8c840176c06946721ef8de53611456e3a81af1951ccd80bed88a4d56ace5b6d"} Oct 02 08:48:46 crc kubenswrapper[4829]: I1002 08:48:46.833613 4829 generic.go:334] "Generic (PLEG): container finished" podID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerID="d8c840176c06946721ef8de53611456e3a81af1951ccd80bed88a4d56ace5b6d" exitCode=0 Oct 02 08:48:46 crc kubenswrapper[4829]: I1002 08:48:46.833732 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerDied","Data":"d8c840176c06946721ef8de53611456e3a81af1951ccd80bed88a4d56ace5b6d"} Oct 02 08:48:47 crc kubenswrapper[4829]: I1002 08:48:47.849745 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerStarted","Data":"d1035b40e2eb1bc4bdb42b3fbecb5f5d54216bb46e4b16624c58e521b5995bdc"} Oct 02 08:48:47 crc kubenswrapper[4829]: I1002 08:48:47.886107 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9mgs5" podStartSLOduration=2.100967798 podStartE2EDuration="7.88607089s" podCreationTimestamp="2025-10-02 08:48:40 +0000 UTC" firstStartedPulling="2025-10-02 08:48:41.775782613 +0000 UTC m=+5513.115431018" lastFinishedPulling="2025-10-02 08:48:47.560885705 +0000 UTC m=+5518.900534110" observedRunningTime="2025-10-02 08:48:47.870639765 +0000 UTC m=+5519.210288180" watchObservedRunningTime="2025-10-02 08:48:47.88607089 +0000 UTC m=+5519.225719335" Oct 02 08:48:50 crc kubenswrapper[4829]: I1002 08:48:50.894346 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:50 crc kubenswrapper[4829]: I1002 08:48:50.895457 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:48:51 crc kubenswrapper[4829]: I1002 08:48:51.972030 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-9mgs5" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="registry-server" probeResult="failure" output=< Oct 02 08:48:51 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:48:51 crc kubenswrapper[4829]: > Oct 02 08:49:00 crc kubenswrapper[4829]: I1002 08:49:00.948448 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:49:01 crc kubenswrapper[4829]: I1002 08:49:01.022965 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:49:01 crc kubenswrapper[4829]: I1002 08:49:01.202551 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9mgs5"] Oct 02 08:49:02 crc kubenswrapper[4829]: I1002 08:49:02.024676 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9mgs5" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="registry-server" containerID="cri-o://d1035b40e2eb1bc4bdb42b3fbecb5f5d54216bb46e4b16624c58e521b5995bdc" gracePeriod=2 Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.037856 4829 generic.go:334] "Generic (PLEG): container finished" podID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerID="d1035b40e2eb1bc4bdb42b3fbecb5f5d54216bb46e4b16624c58e521b5995bdc" exitCode=0 Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.037916 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerDied","Data":"d1035b40e2eb1bc4bdb42b3fbecb5f5d54216bb46e4b16624c58e521b5995bdc"} Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.043474 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mgs5" event={"ID":"df4673d7-4d78-4e00-a610-38afd1c02f87","Type":"ContainerDied","Data":"3f4f8129dc949cdbd7ffaffb62e574fbc91812201564ea261c3a299b0c480cdf"} Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.043511 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f4f8129dc949cdbd7ffaffb62e574fbc91812201564ea261c3a299b0c480cdf" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.255544 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.427460 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-utilities\") pod \"df4673d7-4d78-4e00-a610-38afd1c02f87\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.427593 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-catalog-content\") pod \"df4673d7-4d78-4e00-a610-38afd1c02f87\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.427970 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t274q\" (UniqueName: \"kubernetes.io/projected/df4673d7-4d78-4e00-a610-38afd1c02f87-kube-api-access-t274q\") pod \"df4673d7-4d78-4e00-a610-38afd1c02f87\" (UID: \"df4673d7-4d78-4e00-a610-38afd1c02f87\") " Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.428809 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-utilities" (OuterVolumeSpecName: "utilities") pod "df4673d7-4d78-4e00-a610-38afd1c02f87" (UID: "df4673d7-4d78-4e00-a610-38afd1c02f87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.440416 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df4673d7-4d78-4e00-a610-38afd1c02f87-kube-api-access-t274q" (OuterVolumeSpecName: "kube-api-access-t274q") pod "df4673d7-4d78-4e00-a610-38afd1c02f87" (UID: "df4673d7-4d78-4e00-a610-38afd1c02f87"). InnerVolumeSpecName "kube-api-access-t274q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.503755 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df4673d7-4d78-4e00-a610-38afd1c02f87" (UID: "df4673d7-4d78-4e00-a610-38afd1c02f87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.530862 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t274q\" (UniqueName: \"kubernetes.io/projected/df4673d7-4d78-4e00-a610-38afd1c02f87-kube-api-access-t274q\") on node \"crc\" DevicePath \"\"" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.530891 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:49:03 crc kubenswrapper[4829]: I1002 08:49:03.530900 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4673d7-4d78-4e00-a610-38afd1c02f87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:49:04 crc kubenswrapper[4829]: I1002 08:49:04.055365 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mgs5" Oct 02 08:49:04 crc kubenswrapper[4829]: I1002 08:49:04.110483 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9mgs5"] Oct 02 08:49:04 crc kubenswrapper[4829]: I1002 08:49:04.118785 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9mgs5"] Oct 02 08:49:05 crc kubenswrapper[4829]: I1002 08:49:05.480302 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" path="/var/lib/kubelet/pods/df4673d7-4d78-4e00-a610-38afd1c02f87/volumes" Oct 02 08:49:55 crc kubenswrapper[4829]: I1002 08:49:55.329687 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:49:55 crc kubenswrapper[4829]: I1002 08:49:55.331473 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:50:25 crc kubenswrapper[4829]: I1002 08:50:25.328969 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:50:25 crc kubenswrapper[4829]: I1002 08:50:25.329630 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:50:45 crc kubenswrapper[4829]: I1002 08:50:45.168750 4829 generic.go:334] "Generic (PLEG): container finished" podID="18e0103e-1a6b-4f9c-b583-f3dc8f405137" containerID="7239b6eb333cd5b8373505a0b8edcd4335aa85af66728a195ad41f6dc223ba8b" exitCode=0 Oct 02 08:50:45 crc kubenswrapper[4829]: I1002 08:50:45.168841 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"18e0103e-1a6b-4f9c-b583-f3dc8f405137","Type":"ContainerDied","Data":"7239b6eb333cd5b8373505a0b8edcd4335aa85af66728a195ad41f6dc223ba8b"} Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.585215 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.674939 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ca-certs\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675030 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675083 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ccxm\" (UniqueName: \"kubernetes.io/projected/18e0103e-1a6b-4f9c-b583-f3dc8f405137-kube-api-access-2ccxm\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675105 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ssh-key\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675153 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-config-data\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675277 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-workdir\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675313 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-temporary\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675329 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config-secret\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.675359 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\" (UID: \"18e0103e-1a6b-4f9c-b583-f3dc8f405137\") " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.678990 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.679574 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-config-data" (OuterVolumeSpecName: "config-data") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.681530 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.681952 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18e0103e-1a6b-4f9c-b583-f3dc8f405137-kube-api-access-2ccxm" (OuterVolumeSpecName: "kube-api-access-2ccxm") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "kube-api-access-2ccxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.706049 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.706098 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.714934 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.753357 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.777921 4829 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.777966 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.778001 4829 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.778019 4829 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.778043 4829 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.778096 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ccxm\" (UniqueName: \"kubernetes.io/projected/18e0103e-1a6b-4f9c-b583-f3dc8f405137-kube-api-access-2ccxm\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.778115 4829 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18e0103e-1a6b-4f9c-b583-f3dc8f405137-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.778132 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18e0103e-1a6b-4f9c-b583-f3dc8f405137-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.821330 4829 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 08:50:46 crc kubenswrapper[4829]: I1002 08:50:46.884747 4829 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:47 crc kubenswrapper[4829]: I1002 08:50:47.116866 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "18e0103e-1a6b-4f9c-b583-f3dc8f405137" (UID: "18e0103e-1a6b-4f9c-b583-f3dc8f405137"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:50:47 crc kubenswrapper[4829]: I1002 08:50:47.192619 4829 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/18e0103e-1a6b-4f9c-b583-f3dc8f405137-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 02 08:50:47 crc kubenswrapper[4829]: I1002 08:50:47.194449 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"18e0103e-1a6b-4f9c-b583-f3dc8f405137","Type":"ContainerDied","Data":"622f7d7e84213d616a123ebc9e15c6229565096cbee3023dcf862a90c5f3a58a"} Oct 02 08:50:47 crc kubenswrapper[4829]: I1002 08:50:47.194504 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 02 08:50:47 crc kubenswrapper[4829]: I1002 08:50:47.194507 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="622f7d7e84213d616a123ebc9e15c6229565096cbee3023dcf862a90c5f3a58a" Oct 02 08:50:55 crc kubenswrapper[4829]: I1002 08:50:55.328687 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:50:55 crc kubenswrapper[4829]: I1002 08:50:55.329319 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:50:55 crc kubenswrapper[4829]: I1002 08:50:55.329387 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:50:55 crc kubenswrapper[4829]: I1002 08:50:55.330456 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:50:55 crc kubenswrapper[4829]: I1002 08:50:55.330555 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" gracePeriod=600 Oct 02 08:50:55 crc kubenswrapper[4829]: E1002 08:50:55.474562 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.210483 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 08:50:56 crc kubenswrapper[4829]: E1002 08:50:56.211591 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="extract-content" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.211620 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="extract-content" Oct 02 08:50:56 crc kubenswrapper[4829]: E1002 08:50:56.211646 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="extract-utilities" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.211663 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="extract-utilities" Oct 02 08:50:56 crc kubenswrapper[4829]: E1002 08:50:56.211715 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="registry-server" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.211728 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="registry-server" Oct 02 08:50:56 crc kubenswrapper[4829]: E1002 08:50:56.211758 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18e0103e-1a6b-4f9c-b583-f3dc8f405137" containerName="tempest-tests-tempest-tests-runner" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.211771 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="18e0103e-1a6b-4f9c-b583-f3dc8f405137" containerName="tempest-tests-tempest-tests-runner" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.212110 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="18e0103e-1a6b-4f9c-b583-f3dc8f405137" containerName="tempest-tests-tempest-tests-runner" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.212175 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="df4673d7-4d78-4e00-a610-38afd1c02f87" containerName="registry-server" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.225925 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.228904 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-lsrdp" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.234715 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.292957 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" exitCode=0 Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.293009 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592"} Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.293292 4829 scope.go:117] "RemoveContainer" containerID="fd55a71d4738705ab0f70625de7b8981cb5b16abfb30540fbfe73c64ac1d1a62" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.294524 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:50:56 crc kubenswrapper[4829]: E1002 08:50:56.295002 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.382092 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64svh\" (UniqueName: \"kubernetes.io/projected/f6e01654-3f1b-4049-be7e-a562dfaa59df-kube-api-access-64svh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.382387 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.483830 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64svh\" (UniqueName: \"kubernetes.io/projected/f6e01654-3f1b-4049-be7e-a562dfaa59df-kube-api-access-64svh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.483923 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.484560 4829 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.506978 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64svh\" (UniqueName: \"kubernetes.io/projected/f6e01654-3f1b-4049-be7e-a562dfaa59df-kube-api-access-64svh\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.520304 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f6e01654-3f1b-4049-be7e-a562dfaa59df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:56 crc kubenswrapper[4829]: I1002 08:50:56.560135 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 02 08:50:57 crc kubenswrapper[4829]: I1002 08:50:57.065100 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 02 08:50:57 crc kubenswrapper[4829]: I1002 08:50:57.306696 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f6e01654-3f1b-4049-be7e-a562dfaa59df","Type":"ContainerStarted","Data":"ce19791af03c7f1964a686bf0495cff971fee572918831fdde4940789f909062"} Oct 02 08:50:58 crc kubenswrapper[4829]: I1002 08:50:58.318890 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f6e01654-3f1b-4049-be7e-a562dfaa59df","Type":"ContainerStarted","Data":"04b39a47739a08b65b7ab089ca412a598ae9b8ca9bff3c5464efa27aed7cac18"} Oct 02 08:50:58 crc kubenswrapper[4829]: I1002 08:50:58.350797 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.406016727 podStartE2EDuration="2.350631549s" podCreationTimestamp="2025-10-02 08:50:56 +0000 UTC" firstStartedPulling="2025-10-02 08:50:57.07577072 +0000 UTC m=+5648.415419125" lastFinishedPulling="2025-10-02 08:50:58.020385542 +0000 UTC m=+5649.360033947" observedRunningTime="2025-10-02 08:50:58.33470706 +0000 UTC m=+5649.674355505" watchObservedRunningTime="2025-10-02 08:50:58.350631549 +0000 UTC m=+5649.690416929" Oct 02 08:51:09 crc kubenswrapper[4829]: I1002 08:51:09.474702 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:51:09 crc kubenswrapper[4829]: E1002 08:51:09.475537 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.031757 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5t9l/must-gather-vbb8f"] Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.035859 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.038197 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h5t9l"/"openshift-service-ca.crt" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.038300 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-h5t9l"/"default-dockercfg-dk79k" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.038603 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h5t9l"/"kube-root-ca.crt" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.048925 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h5t9l/must-gather-vbb8f"] Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.147708 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snd7c\" (UniqueName: \"kubernetes.io/projected/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-kube-api-access-snd7c\") pod \"must-gather-vbb8f\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.147974 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-must-gather-output\") pod \"must-gather-vbb8f\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.250007 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snd7c\" (UniqueName: \"kubernetes.io/projected/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-kube-api-access-snd7c\") pod \"must-gather-vbb8f\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.250071 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-must-gather-output\") pod \"must-gather-vbb8f\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.250887 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-must-gather-output\") pod \"must-gather-vbb8f\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.277084 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snd7c\" (UniqueName: \"kubernetes.io/projected/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-kube-api-access-snd7c\") pod \"must-gather-vbb8f\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.357487 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:51:16 crc kubenswrapper[4829]: I1002 08:51:16.859943 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h5t9l/must-gather-vbb8f"] Oct 02 08:51:17 crc kubenswrapper[4829]: I1002 08:51:17.574001 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" event={"ID":"b1c62e6e-1e8e-4015-8b54-4179909c1c9b","Type":"ContainerStarted","Data":"2928f8e21f1852dbb0fa2a1f6578f788fffa26e7db89182033ba36d0baca0fac"} Oct 02 08:51:23 crc kubenswrapper[4829]: I1002 08:51:23.466017 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:51:23 crc kubenswrapper[4829]: E1002 08:51:23.466830 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:51:23 crc kubenswrapper[4829]: I1002 08:51:23.631432 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" event={"ID":"b1c62e6e-1e8e-4015-8b54-4179909c1c9b","Type":"ContainerStarted","Data":"3e6112ec904a756317da0aeec6c284d41482e67647573e20b7f40f333aea4534"} Oct 02 08:51:24 crc kubenswrapper[4829]: I1002 08:51:24.658135 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" event={"ID":"b1c62e6e-1e8e-4015-8b54-4179909c1c9b","Type":"ContainerStarted","Data":"22dcb65ed5b103d53342b86445aca363fde7b72218ba220c97ac62c5fb360787"} Oct 02 08:51:24 crc kubenswrapper[4829]: I1002 08:51:24.682185 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" podStartSLOduration=2.226108124 podStartE2EDuration="8.682154014s" podCreationTimestamp="2025-10-02 08:51:16 +0000 UTC" firstStartedPulling="2025-10-02 08:51:16.904817056 +0000 UTC m=+5668.244465461" lastFinishedPulling="2025-10-02 08:51:23.360862916 +0000 UTC m=+5674.700511351" observedRunningTime="2025-10-02 08:51:24.674580306 +0000 UTC m=+5676.014228751" watchObservedRunningTime="2025-10-02 08:51:24.682154014 +0000 UTC m=+5676.021802429" Oct 02 08:51:27 crc kubenswrapper[4829]: E1002 08:51:27.626580 4829 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:43912->38.102.83.200:43623: write tcp 38.102.83.200:43912->38.102.83.200:43623: write: broken pipe Oct 02 08:51:27 crc kubenswrapper[4829]: E1002 08:51:27.714948 4829 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.200:43948->38.102.83.200:43623: read tcp 38.102.83.200:43948->38.102.83.200:43623: read: connection reset by peer Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.286330 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-qjpgg"] Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.288328 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.414958 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a74d5ac-8430-4e3d-8e69-b416fca59164-host\") pod \"crc-debug-qjpgg\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.415110 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27rdb\" (UniqueName: \"kubernetes.io/projected/3a74d5ac-8430-4e3d-8e69-b416fca59164-kube-api-access-27rdb\") pod \"crc-debug-qjpgg\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.517344 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a74d5ac-8430-4e3d-8e69-b416fca59164-host\") pod \"crc-debug-qjpgg\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.517512 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27rdb\" (UniqueName: \"kubernetes.io/projected/3a74d5ac-8430-4e3d-8e69-b416fca59164-kube-api-access-27rdb\") pod \"crc-debug-qjpgg\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.517571 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a74d5ac-8430-4e3d-8e69-b416fca59164-host\") pod \"crc-debug-qjpgg\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.699912 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27rdb\" (UniqueName: \"kubernetes.io/projected/3a74d5ac-8430-4e3d-8e69-b416fca59164-kube-api-access-27rdb\") pod \"crc-debug-qjpgg\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: I1002 08:51:28.906409 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:51:28 crc kubenswrapper[4829]: W1002 08:51:28.938361 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a74d5ac_8430_4e3d_8e69_b416fca59164.slice/crio-d0eb98d0f87e7e3177f3e4ef9654375eb7f4831bd6b5d49031203f88e9b651cc WatchSource:0}: Error finding container d0eb98d0f87e7e3177f3e4ef9654375eb7f4831bd6b5d49031203f88e9b651cc: Status 404 returned error can't find the container with id d0eb98d0f87e7e3177f3e4ef9654375eb7f4831bd6b5d49031203f88e9b651cc Oct 02 08:51:29 crc kubenswrapper[4829]: I1002 08:51:29.712216 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" event={"ID":"3a74d5ac-8430-4e3d-8e69-b416fca59164","Type":"ContainerStarted","Data":"d0eb98d0f87e7e3177f3e4ef9654375eb7f4831bd6b5d49031203f88e9b651cc"} Oct 02 08:51:36 crc kubenswrapper[4829]: I1002 08:51:36.461649 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:51:36 crc kubenswrapper[4829]: E1002 08:51:36.462469 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:51:40 crc kubenswrapper[4829]: I1002 08:51:40.836891 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" event={"ID":"3a74d5ac-8430-4e3d-8e69-b416fca59164","Type":"ContainerStarted","Data":"189d0151ed4bbe9c076609980f9a54a72f9c8220f3800d74273079d30ff766b6"} Oct 02 08:51:40 crc kubenswrapper[4829]: I1002 08:51:40.857984 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" podStartSLOduration=1.887465752 podStartE2EDuration="12.857970622s" podCreationTimestamp="2025-10-02 08:51:28 +0000 UTC" firstStartedPulling="2025-10-02 08:51:28.943025865 +0000 UTC m=+5680.282674280" lastFinishedPulling="2025-10-02 08:51:39.913530745 +0000 UTC m=+5691.253179150" observedRunningTime="2025-10-02 08:51:40.851525488 +0000 UTC m=+5692.191173893" watchObservedRunningTime="2025-10-02 08:51:40.857970622 +0000 UTC m=+5692.197619027" Oct 02 08:51:51 crc kubenswrapper[4829]: I1002 08:51:51.461286 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:51:51 crc kubenswrapper[4829]: E1002 08:51:51.462029 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:52:03 crc kubenswrapper[4829]: I1002 08:52:03.461371 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:52:03 crc kubenswrapper[4829]: E1002 08:52:03.462382 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.050179 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fqzkx"] Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.052451 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.066665 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqzkx"] Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.172680 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5c9m\" (UniqueName: \"kubernetes.io/projected/38091a94-8cbf-4d3b-81e6-0651d12166b2-kube-api-access-l5c9m\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.172948 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-utilities\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.173005 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-catalog-content\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.275340 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5c9m\" (UniqueName: \"kubernetes.io/projected/38091a94-8cbf-4d3b-81e6-0651d12166b2-kube-api-access-l5c9m\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.275398 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-utilities\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.275452 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-catalog-content\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.275980 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-catalog-content\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.276073 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-utilities\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.293980 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5c9m\" (UniqueName: \"kubernetes.io/projected/38091a94-8cbf-4d3b-81e6-0651d12166b2-kube-api-access-l5c9m\") pod \"certified-operators-fqzkx\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.374188 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:15 crc kubenswrapper[4829]: I1002 08:52:15.965859 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fqzkx"] Oct 02 08:52:16 crc kubenswrapper[4829]: I1002 08:52:16.176819 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerStarted","Data":"5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846"} Oct 02 08:52:16 crc kubenswrapper[4829]: I1002 08:52:16.177207 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerStarted","Data":"2f5940b9da20dbc3c894b61b46411f32112e937f8e591208ffa33274d637794b"} Oct 02 08:52:17 crc kubenswrapper[4829]: I1002 08:52:17.188207 4829 generic.go:334] "Generic (PLEG): container finished" podID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerID="5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846" exitCode=0 Oct 02 08:52:17 crc kubenswrapper[4829]: I1002 08:52:17.188397 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerDied","Data":"5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846"} Oct 02 08:52:17 crc kubenswrapper[4829]: I1002 08:52:17.465348 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:52:17 crc kubenswrapper[4829]: E1002 08:52:17.465923 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:52:18 crc kubenswrapper[4829]: I1002 08:52:18.200215 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerStarted","Data":"5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad"} Oct 02 08:52:20 crc kubenswrapper[4829]: I1002 08:52:20.220810 4829 generic.go:334] "Generic (PLEG): container finished" podID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerID="5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad" exitCode=0 Oct 02 08:52:20 crc kubenswrapper[4829]: I1002 08:52:20.220914 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerDied","Data":"5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad"} Oct 02 08:52:21 crc kubenswrapper[4829]: I1002 08:52:21.233722 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerStarted","Data":"790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571"} Oct 02 08:52:21 crc kubenswrapper[4829]: I1002 08:52:21.255560 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fqzkx" podStartSLOduration=2.657921914 podStartE2EDuration="6.255545548s" podCreationTimestamp="2025-10-02 08:52:15 +0000 UTC" firstStartedPulling="2025-10-02 08:52:17.193297329 +0000 UTC m=+5728.532945734" lastFinishedPulling="2025-10-02 08:52:20.790920953 +0000 UTC m=+5732.130569368" observedRunningTime="2025-10-02 08:52:21.253055342 +0000 UTC m=+5732.592703747" watchObservedRunningTime="2025-10-02 08:52:21.255545548 +0000 UTC m=+5732.595193953" Oct 02 08:52:25 crc kubenswrapper[4829]: I1002 08:52:25.374524 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:25 crc kubenswrapper[4829]: I1002 08:52:25.375257 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:25 crc kubenswrapper[4829]: I1002 08:52:25.428247 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:26 crc kubenswrapper[4829]: I1002 08:52:26.331695 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:26 crc kubenswrapper[4829]: I1002 08:52:26.382362 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqzkx"] Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.291007 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fqzkx" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="registry-server" containerID="cri-o://790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571" gracePeriod=2 Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.809694 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.954326 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5c9m\" (UniqueName: \"kubernetes.io/projected/38091a94-8cbf-4d3b-81e6-0651d12166b2-kube-api-access-l5c9m\") pod \"38091a94-8cbf-4d3b-81e6-0651d12166b2\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.954410 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-catalog-content\") pod \"38091a94-8cbf-4d3b-81e6-0651d12166b2\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.954553 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-utilities\") pod \"38091a94-8cbf-4d3b-81e6-0651d12166b2\" (UID: \"38091a94-8cbf-4d3b-81e6-0651d12166b2\") " Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.955295 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-utilities" (OuterVolumeSpecName: "utilities") pod "38091a94-8cbf-4d3b-81e6-0651d12166b2" (UID: "38091a94-8cbf-4d3b-81e6-0651d12166b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:52:28 crc kubenswrapper[4829]: I1002 08:52:28.979893 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38091a94-8cbf-4d3b-81e6-0651d12166b2-kube-api-access-l5c9m" (OuterVolumeSpecName: "kube-api-access-l5c9m") pod "38091a94-8cbf-4d3b-81e6-0651d12166b2" (UID: "38091a94-8cbf-4d3b-81e6-0651d12166b2"). InnerVolumeSpecName "kube-api-access-l5c9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.009414 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38091a94-8cbf-4d3b-81e6-0651d12166b2" (UID: "38091a94-8cbf-4d3b-81e6-0651d12166b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.056902 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.057112 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5c9m\" (UniqueName: \"kubernetes.io/projected/38091a94-8cbf-4d3b-81e6-0651d12166b2-kube-api-access-l5c9m\") on node \"crc\" DevicePath \"\"" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.057172 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38091a94-8cbf-4d3b-81e6-0651d12166b2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.300823 4829 generic.go:334] "Generic (PLEG): container finished" podID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerID="790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571" exitCode=0 Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.300862 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerDied","Data":"790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571"} Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.300885 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fqzkx" event={"ID":"38091a94-8cbf-4d3b-81e6-0651d12166b2","Type":"ContainerDied","Data":"2f5940b9da20dbc3c894b61b46411f32112e937f8e591208ffa33274d637794b"} Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.300904 4829 scope.go:117] "RemoveContainer" containerID="790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.301030 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fqzkx" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.330603 4829 scope.go:117] "RemoveContainer" containerID="5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.337139 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fqzkx"] Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.362603 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fqzkx"] Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.369451 4829 scope.go:117] "RemoveContainer" containerID="5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.430909 4829 scope.go:117] "RemoveContainer" containerID="790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571" Oct 02 08:52:29 crc kubenswrapper[4829]: E1002 08:52:29.431310 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571\": container with ID starting with 790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571 not found: ID does not exist" containerID="790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.431351 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571"} err="failed to get container status \"790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571\": rpc error: code = NotFound desc = could not find container \"790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571\": container with ID starting with 790ec85849547fc3bb861bac85dc37484d3f1a41b64ae156c4841e74fb76c571 not found: ID does not exist" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.431377 4829 scope.go:117] "RemoveContainer" containerID="5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad" Oct 02 08:52:29 crc kubenswrapper[4829]: E1002 08:52:29.431701 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad\": container with ID starting with 5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad not found: ID does not exist" containerID="5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.431743 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad"} err="failed to get container status \"5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad\": rpc error: code = NotFound desc = could not find container \"5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad\": container with ID starting with 5654df25ed1985ec6e9b37ead7fc49e590aa791d79af8a16e21d91b4f8a9fbad not found: ID does not exist" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.431770 4829 scope.go:117] "RemoveContainer" containerID="5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846" Oct 02 08:52:29 crc kubenswrapper[4829]: E1002 08:52:29.431988 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846\": container with ID starting with 5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846 not found: ID does not exist" containerID="5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.432008 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846"} err="failed to get container status \"5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846\": rpc error: code = NotFound desc = could not find container \"5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846\": container with ID starting with 5faeaa1992309709fb58b7ab50278df84937f22b1eec93fb446add5eebfe3846 not found: ID does not exist" Oct 02 08:52:29 crc kubenswrapper[4829]: I1002 08:52:29.474551 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" path="/var/lib/kubelet/pods/38091a94-8cbf-4d3b-81e6-0651d12166b2/volumes" Oct 02 08:52:31 crc kubenswrapper[4829]: I1002 08:52:31.465478 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:52:31 crc kubenswrapper[4829]: E1002 08:52:31.466334 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:52:42 crc kubenswrapper[4829]: I1002 08:52:42.462169 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:52:42 crc kubenswrapper[4829]: E1002 08:52:42.462923 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:52:51 crc kubenswrapper[4829]: I1002 08:52:51.845885 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7944ccd6-g5gh7_69f201db-52b1-4a9f-a1f3-adb1bdb923da/barbican-api/0.log" Oct 02 08:52:51 crc kubenswrapper[4829]: I1002 08:52:51.963663 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7944ccd6-g5gh7_69f201db-52b1-4a9f-a1f3-adb1bdb923da/barbican-api-log/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.096717 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7cf66b5f98-d625p_1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244/barbican-keystone-listener/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.201157 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7cf66b5f98-d625p_1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244/barbican-keystone-listener-log/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.257951 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f9d8b6c5-ds24l_54708879-c00e-4582-a37e-5becf429dc5d/barbican-worker/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.448887 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f9d8b6c5-ds24l_54708879-c00e-4582-a37e-5becf429dc5d/barbican-worker-log/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.535960 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk_46cccbf0-9925-4708-bd05-119652ca2732/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.692030 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/ceilometer-central-agent/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.724396 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/ceilometer-notification-agent/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.928613 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/proxy-httpd/0.log" Oct 02 08:52:52 crc kubenswrapper[4829]: I1002 08:52:52.943914 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/sg-core/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.275998 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2bec1c3c-60b9-4776-93c0-3d51886b3552/cinder-api-log/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.289528 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2bec1c3c-60b9-4776-93c0-3d51886b3552/cinder-api/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.432868 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5da522c4-7a44-46bc-b84e-eb230aad0bb7/cinder-scheduler/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.513767 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5da522c4-7a44-46bc-b84e-eb230aad0bb7/probe/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.648025 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz_e6b2c2b0-9a68-4525-9f18-64b8e4354246/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.750040 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k746m_28b94275-3752-42ab-b862-fee2ee0ead47/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:53 crc kubenswrapper[4829]: I1002 08:52:53.906951 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m_f418c44c-86bc-4ba6-aabb-da19658c0e77/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.130907 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-cvpg4_0c1b029b-bf2c-4b4b-b71c-6050f640212b/init/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.340344 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-cvpg4_0c1b029b-bf2c-4b4b-b71c-6050f640212b/init/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.444277 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-cvpg4_0c1b029b-bf2c-4b4b-b71c-6050f640212b/dnsmasq-dns/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.519791 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp_ac409859-eb90-422d-ad2f-4cf60f171844/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.646794 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3f017e85-e428-48e7-9b00-705ca534019d/glance-httpd/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.713078 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3f017e85-e428-48e7-9b00-705ca534019d/glance-log/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.818940 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_9c3dbee2-4ec4-4d14-8e07-8a04428f0429/glance-httpd/0.log" Oct 02 08:52:54 crc kubenswrapper[4829]: I1002 08:52:54.896407 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_9c3dbee2-4ec4-4d14-8e07-8a04428f0429/glance-log/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.099835 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5df99f8898-76z9s_3407e6aa-fc06-4a2e-bd97-b8540ae12167/horizon/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.120162 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-7krkz_28912442-1f56-461e-9d26-18227d1f2386/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.327711 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-dfmsw_4181a86c-75a3-4400-9b89-352c04c06044/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.601192 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323201-j6v6q_0ca1b8dd-db65-4bcf-908c-3bf2b145d334/keystone-cron/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.776462 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_fadeaa43-1547-4915-9e34-43e880485d84/kube-state-metrics/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.789313 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5df99f8898-76z9s_3407e6aa-fc06-4a2e-bd97-b8540ae12167/horizon-log/0.log" Oct 02 08:52:55 crc kubenswrapper[4829]: I1002 08:52:55.993383 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-94ftd_9dfa4494-c2fb-415c-a38a-79afbf13be81/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:56 crc kubenswrapper[4829]: I1002 08:52:56.068899 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6b77fd5f6d-mmltx_5d8b5f01-0374-4315-82d6-de60c41b6bad/keystone-api/0.log" Oct 02 08:52:56 crc kubenswrapper[4829]: I1002 08:52:56.460650 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:52:56 crc kubenswrapper[4829]: E1002 08:52:56.461793 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:52:56 crc kubenswrapper[4829]: I1002 08:52:56.563765 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79b8d665c5-dgn5f_1a40a5fd-a82e-430c-a15d-e32f0a14f5f2/neutron-httpd/0.log" Oct 02 08:52:56 crc kubenswrapper[4829]: I1002 08:52:56.618081 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79b8d665c5-dgn5f_1a40a5fd-a82e-430c-a15d-e32f0a14f5f2/neutron-api/0.log" Oct 02 08:52:56 crc kubenswrapper[4829]: I1002 08:52:56.666723 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc_d0041618-1a4e-474f-8570-ccf383c6cff9/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:57 crc kubenswrapper[4829]: I1002 08:52:57.479327 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_911b568c-7273-4d79-8213-0d2577e98046/nova-cell0-conductor-conductor/0.log" Oct 02 08:52:58 crc kubenswrapper[4829]: I1002 08:52:58.039766 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_483a24aa-87d3-4b0a-accd-5ca4a8a7a029/nova-cell1-conductor-conductor/0.log" Oct 02 08:52:58 crc kubenswrapper[4829]: I1002 08:52:58.098065 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23175b6b-8b9a-43da-a7ba-d0b5cb04863a/nova-api-log/0.log" Oct 02 08:52:58 crc kubenswrapper[4829]: I1002 08:52:58.449803 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_23a2af5c-929f-41a7-8a96-d42f5902f200/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 08:52:58 crc kubenswrapper[4829]: I1002 08:52:58.494194 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23175b6b-8b9a-43da-a7ba-d0b5cb04863a/nova-api-api/0.log" Oct 02 08:52:58 crc kubenswrapper[4829]: I1002 08:52:58.779903 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1298703b-f77d-4cf5-9537-8928493b4e6b/nova-metadata-log/0.log" Oct 02 08:52:58 crc kubenswrapper[4829]: I1002 08:52:58.794062 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jmb99_97915d50-01b1-415e-99f1-17c7d4340c66/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:52:59 crc kubenswrapper[4829]: I1002 08:52:59.356735 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_52f60623-b76f-43dc-9022-d62914e8950b/nova-scheduler-scheduler/0.log" Oct 02 08:52:59 crc kubenswrapper[4829]: I1002 08:52:59.464850 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d72fc392-aa3b-40de-98cd-a0f7a633d5fa/mysql-bootstrap/0.log" Oct 02 08:52:59 crc kubenswrapper[4829]: I1002 08:52:59.694322 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d72fc392-aa3b-40de-98cd-a0f7a633d5fa/mysql-bootstrap/0.log" Oct 02 08:52:59 crc kubenswrapper[4829]: I1002 08:52:59.708172 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d72fc392-aa3b-40de-98cd-a0f7a633d5fa/galera/0.log" Oct 02 08:52:59 crc kubenswrapper[4829]: I1002 08:52:59.903240 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0195e48a-f8c9-433b-a9a8-c3da38457280/mysql-bootstrap/0.log" Oct 02 08:53:00 crc kubenswrapper[4829]: I1002 08:53:00.132892 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0195e48a-f8c9-433b-a9a8-c3da38457280/galera/0.log" Oct 02 08:53:00 crc kubenswrapper[4829]: I1002 08:53:00.164444 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0195e48a-f8c9-433b-a9a8-c3da38457280/mysql-bootstrap/0.log" Oct 02 08:53:00 crc kubenswrapper[4829]: I1002 08:53:00.430646 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_859798c1-faae-46b4-83f2-0b9d18734236/openstackclient/0.log" Oct 02 08:53:00 crc kubenswrapper[4829]: I1002 08:53:00.620138 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-s5fpj_4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3/openstack-network-exporter/0.log" Oct 02 08:53:00 crc kubenswrapper[4829]: I1002 08:53:00.818129 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-njw55_91ff1e5f-73e2-4237-b201-84065a586553/ovn-controller/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.002684 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1298703b-f77d-4cf5-9537-8928493b4e6b/nova-metadata-metadata/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.076215 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovsdb-server-init/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.207285 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovsdb-server-init/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.259471 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovs-vswitchd/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.275601 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovsdb-server/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.462382 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xsv2v_adb0b737-9122-4893-8b4d-2111d18c2a4d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.643211 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_775c00d4-e88e-42a8-9351-edd61e0c2cd6/openstack-network-exporter/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.690265 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_775c00d4-e88e-42a8-9351-edd61e0c2cd6/ovn-northd/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.841667 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_37f64531-608d-4dd5-99ad-5c928bcf7ca0/openstack-network-exporter/0.log" Oct 02 08:53:01 crc kubenswrapper[4829]: I1002 08:53:01.879934 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_37f64531-608d-4dd5-99ad-5c928bcf7ca0/ovsdbserver-nb/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.046587 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cedf4e9a-cfe0-4499-8600-edf90b9b5cda/openstack-network-exporter/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.053373 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cedf4e9a-cfe0-4499-8600-edf90b9b5cda/ovsdbserver-sb/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.419212 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-995ff44cb-bcb2z_f49ffa44-6b66-43b1-9201-afc0e8bb9311/placement-api/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.542176 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-995ff44cb-bcb2z_f49ffa44-6b66-43b1-9201-afc0e8bb9311/placement-log/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.622476 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/init-config-reloader/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.775281 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/init-config-reloader/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.809199 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/prometheus/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.819725 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/config-reloader/0.log" Oct 02 08:53:02 crc kubenswrapper[4829]: I1002 08:53:02.962140 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/thanos-sidecar/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.015014 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66249bb-0358-455a-8ca1-7c9688cfe5f0/setup-container/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.171185 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66249bb-0358-455a-8ca1-7c9688cfe5f0/setup-container/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.193435 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66249bb-0358-455a-8ca1-7c9688cfe5f0/rabbitmq/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.331682 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_43522a1c-881d-4d33-91e3-3dea8c241d8c/setup-container/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.530588 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_43522a1c-881d-4d33-91e3-3dea8c241d8c/rabbitmq/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.553068 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_43522a1c-881d-4d33-91e3-3dea8c241d8c/setup-container/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.723766 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87_4d515775-5135-4f3f-aa21-fd1f648afb11/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.780327 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fm777_fb03e152-ef79-4d7f-920d-5d1fbacb7167/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:03 crc kubenswrapper[4829]: I1002 08:53:03.950061 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26_01ec7e0b-c62e-4bd4-ab7e-6756c5314161/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.190092 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2kgpj_3ed1e719-596e-42fd-ac3f-b092996c9ca6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.265647 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8b6wr_0506e251-5b72-4ce4-91c5-8116bf387c34/ssh-known-hosts-edpm-deployment/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.471280 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b89d7c8c-f2lgb_395b490d-fbc2-403d-ac35-9541329363c2/proxy-server/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.674629 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b89d7c8c-f2lgb_395b490d-fbc2-403d-ac35-9541329363c2/proxy-httpd/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.694849 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m6p2d_b5063471-4946-4763-9338-1ad3fbb4b9ef/swift-ring-rebalance/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.866281 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-auditor/0.log" Oct 02 08:53:04 crc kubenswrapper[4829]: I1002 08:53:04.950598 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-reaper/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.115622 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-replicator/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.147898 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-auditor/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.179467 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-server/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.342960 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-server/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.348096 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-replicator/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.370550 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-updater/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.526182 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-expirer/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.554516 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-auditor/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.604369 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-replicator/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.760421 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-updater/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.784066 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-server/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.831250 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/rsync/0.log" Oct 02 08:53:05 crc kubenswrapper[4829]: I1002 08:53:05.951213 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/swift-recon-cron/0.log" Oct 02 08:53:06 crc kubenswrapper[4829]: I1002 08:53:06.167455 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx_e74d0c61-084d-49d1-877a-3325875d5c15/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:06 crc kubenswrapper[4829]: I1002 08:53:06.301694 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_18e0103e-1a6b-4f9c-b583-f3dc8f405137/tempest-tests-tempest-tests-runner/0.log" Oct 02 08:53:06 crc kubenswrapper[4829]: I1002 08:53:06.407283 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f6e01654-3f1b-4049-be7e-a562dfaa59df/test-operator-logs-container/0.log" Oct 02 08:53:06 crc kubenswrapper[4829]: I1002 08:53:06.598883 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5_bb87d295-2bab-44ea-b5cc-cc77a0efe964/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 08:53:07 crc kubenswrapper[4829]: I1002 08:53:07.461587 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:53:07 crc kubenswrapper[4829]: E1002 08:53:07.461963 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:53:07 crc kubenswrapper[4829]: I1002 08:53:07.497768 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_59d8c82e-9aca-41ca-9d4b-8f5ec2824132/watcher-applier/0.log" Oct 02 08:53:07 crc kubenswrapper[4829]: I1002 08:53:07.695938 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_ab94865b-0ce8-4b10-bbab-7354e3603e1b/watcher-api-log/0.log" Oct 02 08:53:09 crc kubenswrapper[4829]: I1002 08:53:09.123830 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_c6390f00-cfcd-4b1c-831d-fd355daac37e/watcher-decision-engine/0.log" Oct 02 08:53:11 crc kubenswrapper[4829]: I1002 08:53:11.497289 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_ab94865b-0ce8-4b10-bbab-7354e3603e1b/watcher-api/0.log" Oct 02 08:53:14 crc kubenswrapper[4829]: I1002 08:53:14.018414 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5146035b-0eb1-40c3-bc13-9d40affa0f2e/memcached/0.log" Oct 02 08:53:20 crc kubenswrapper[4829]: I1002 08:53:20.461610 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:53:20 crc kubenswrapper[4829]: E1002 08:53:20.462463 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:53:35 crc kubenswrapper[4829]: I1002 08:53:35.461448 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:53:35 crc kubenswrapper[4829]: E1002 08:53:35.462822 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.029890 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6p926"] Oct 02 08:53:38 crc kubenswrapper[4829]: E1002 08:53:38.030864 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="extract-content" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.030897 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="extract-content" Oct 02 08:53:38 crc kubenswrapper[4829]: E1002 08:53:38.030935 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="extract-utilities" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.030948 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="extract-utilities" Oct 02 08:53:38 crc kubenswrapper[4829]: E1002 08:53:38.030991 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="registry-server" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.031003 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="registry-server" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.031351 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="38091a94-8cbf-4d3b-81e6-0651d12166b2" containerName="registry-server" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.033527 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.050362 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6p926"] Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.178356 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-catalog-content\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.178446 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-utilities\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.178472 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnbts\" (UniqueName: \"kubernetes.io/projected/b4a659f4-409d-434c-b3de-ef23329bf386-kube-api-access-fnbts\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.280876 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-catalog-content\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.281006 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-utilities\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.281815 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-utilities\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.281039 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnbts\" (UniqueName: \"kubernetes.io/projected/b4a659f4-409d-434c-b3de-ef23329bf386-kube-api-access-fnbts\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.281900 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-catalog-content\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.304827 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnbts\" (UniqueName: \"kubernetes.io/projected/b4a659f4-409d-434c-b3de-ef23329bf386-kube-api-access-fnbts\") pod \"redhat-operators-6p926\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.363506 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:38 crc kubenswrapper[4829]: I1002 08:53:38.702051 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6p926"] Oct 02 08:53:39 crc kubenswrapper[4829]: I1002 08:53:39.059471 4829 generic.go:334] "Generic (PLEG): container finished" podID="b4a659f4-409d-434c-b3de-ef23329bf386" containerID="cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2" exitCode=0 Oct 02 08:53:39 crc kubenswrapper[4829]: I1002 08:53:39.059512 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerDied","Data":"cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2"} Oct 02 08:53:39 crc kubenswrapper[4829]: I1002 08:53:39.059539 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerStarted","Data":"d153a449000ba99879c77dcba4bfc1ebabe64ad0f43c508828916102393377c4"} Oct 02 08:53:39 crc kubenswrapper[4829]: I1002 08:53:39.061021 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:53:41 crc kubenswrapper[4829]: I1002 08:53:41.088292 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerStarted","Data":"ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54"} Oct 02 08:53:43 crc kubenswrapper[4829]: I1002 08:53:43.121690 4829 generic.go:334] "Generic (PLEG): container finished" podID="b4a659f4-409d-434c-b3de-ef23329bf386" containerID="ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54" exitCode=0 Oct 02 08:53:43 crc kubenswrapper[4829]: I1002 08:53:43.121766 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerDied","Data":"ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54"} Oct 02 08:53:44 crc kubenswrapper[4829]: I1002 08:53:44.139888 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerStarted","Data":"c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331"} Oct 02 08:53:44 crc kubenswrapper[4829]: I1002 08:53:44.173777 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6p926" podStartSLOduration=2.590425114 podStartE2EDuration="7.173759254s" podCreationTimestamp="2025-10-02 08:53:37 +0000 UTC" firstStartedPulling="2025-10-02 08:53:39.060748789 +0000 UTC m=+5810.400397194" lastFinishedPulling="2025-10-02 08:53:43.644082889 +0000 UTC m=+5814.983731334" observedRunningTime="2025-10-02 08:53:44.169841572 +0000 UTC m=+5815.509490047" watchObservedRunningTime="2025-10-02 08:53:44.173759254 +0000 UTC m=+5815.513407659" Oct 02 08:53:48 crc kubenswrapper[4829]: I1002 08:53:48.192140 4829 generic.go:334] "Generic (PLEG): container finished" podID="3a74d5ac-8430-4e3d-8e69-b416fca59164" containerID="189d0151ed4bbe9c076609980f9a54a72f9c8220f3800d74273079d30ff766b6" exitCode=0 Oct 02 08:53:48 crc kubenswrapper[4829]: I1002 08:53:48.192221 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" event={"ID":"3a74d5ac-8430-4e3d-8e69-b416fca59164","Type":"ContainerDied","Data":"189d0151ed4bbe9c076609980f9a54a72f9c8220f3800d74273079d30ff766b6"} Oct 02 08:53:48 crc kubenswrapper[4829]: I1002 08:53:48.364460 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:48 crc kubenswrapper[4829]: I1002 08:53:48.364519 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.316309 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.373358 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-qjpgg"] Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.386266 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-qjpgg"] Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.436067 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27rdb\" (UniqueName: \"kubernetes.io/projected/3a74d5ac-8430-4e3d-8e69-b416fca59164-kube-api-access-27rdb\") pod \"3a74d5ac-8430-4e3d-8e69-b416fca59164\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.436548 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a74d5ac-8430-4e3d-8e69-b416fca59164-host\") pod \"3a74d5ac-8430-4e3d-8e69-b416fca59164\" (UID: \"3a74d5ac-8430-4e3d-8e69-b416fca59164\") " Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.436651 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a74d5ac-8430-4e3d-8e69-b416fca59164-host" (OuterVolumeSpecName: "host") pod "3a74d5ac-8430-4e3d-8e69-b416fca59164" (UID: "3a74d5ac-8430-4e3d-8e69-b416fca59164"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.437527 4829 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3a74d5ac-8430-4e3d-8e69-b416fca59164-host\") on node \"crc\" DevicePath \"\"" Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.445010 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a74d5ac-8430-4e3d-8e69-b416fca59164-kube-api-access-27rdb" (OuterVolumeSpecName: "kube-api-access-27rdb") pod "3a74d5ac-8430-4e3d-8e69-b416fca59164" (UID: "3a74d5ac-8430-4e3d-8e69-b416fca59164"). InnerVolumeSpecName "kube-api-access-27rdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.453010 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6p926" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="registry-server" probeResult="failure" output=< Oct 02 08:53:49 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 08:53:49 crc kubenswrapper[4829]: > Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.484092 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a74d5ac-8430-4e3d-8e69-b416fca59164" path="/var/lib/kubelet/pods/3a74d5ac-8430-4e3d-8e69-b416fca59164/volumes" Oct 02 08:53:49 crc kubenswrapper[4829]: I1002 08:53:49.539133 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27rdb\" (UniqueName: \"kubernetes.io/projected/3a74d5ac-8430-4e3d-8e69-b416fca59164-kube-api-access-27rdb\") on node \"crc\" DevicePath \"\"" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.218511 4829 scope.go:117] "RemoveContainer" containerID="189d0151ed4bbe9c076609980f9a54a72f9c8220f3800d74273079d30ff766b6" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.218596 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-qjpgg" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.461607 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:53:50 crc kubenswrapper[4829]: E1002 08:53:50.462187 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.545160 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-xv8hz"] Oct 02 08:53:50 crc kubenswrapper[4829]: E1002 08:53:50.546164 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a74d5ac-8430-4e3d-8e69-b416fca59164" containerName="container-00" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.546191 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a74d5ac-8430-4e3d-8e69-b416fca59164" containerName="container-00" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.546632 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a74d5ac-8430-4e3d-8e69-b416fca59164" containerName="container-00" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.547413 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.663636 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8gf7\" (UniqueName: \"kubernetes.io/projected/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-kube-api-access-h8gf7\") pod \"crc-debug-xv8hz\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.663676 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-host\") pod \"crc-debug-xv8hz\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.767004 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8gf7\" (UniqueName: \"kubernetes.io/projected/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-kube-api-access-h8gf7\") pod \"crc-debug-xv8hz\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.767054 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-host\") pod \"crc-debug-xv8hz\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.767321 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-host\") pod \"crc-debug-xv8hz\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.792413 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8gf7\" (UniqueName: \"kubernetes.io/projected/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-kube-api-access-h8gf7\") pod \"crc-debug-xv8hz\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:50 crc kubenswrapper[4829]: I1002 08:53:50.866485 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:51 crc kubenswrapper[4829]: I1002 08:53:51.232969 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" event={"ID":"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0","Type":"ContainerStarted","Data":"cc3ead9b692ac4dc301248b000e155f1ae4fd243c198c5b4648aa8ede7c60e51"} Oct 02 08:53:51 crc kubenswrapper[4829]: I1002 08:53:51.233021 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" event={"ID":"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0","Type":"ContainerStarted","Data":"e59ff5aa9d5c69084cc173e351a931a2673655636396e8b938350e64a57a24fd"} Oct 02 08:53:51 crc kubenswrapper[4829]: I1002 08:53:51.271054 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" podStartSLOduration=1.271028307 podStartE2EDuration="1.271028307s" podCreationTimestamp="2025-10-02 08:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:53:51.246899584 +0000 UTC m=+5822.586548019" watchObservedRunningTime="2025-10-02 08:53:51.271028307 +0000 UTC m=+5822.610676742" Oct 02 08:53:53 crc kubenswrapper[4829]: I1002 08:53:53.253325 4829 generic.go:334] "Generic (PLEG): container finished" podID="1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" containerID="cc3ead9b692ac4dc301248b000e155f1ae4fd243c198c5b4648aa8ede7c60e51" exitCode=0 Oct 02 08:53:53 crc kubenswrapper[4829]: I1002 08:53:53.253390 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" event={"ID":"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0","Type":"ContainerDied","Data":"cc3ead9b692ac4dc301248b000e155f1ae4fd243c198c5b4648aa8ede7c60e51"} Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.360109 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.454699 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8gf7\" (UniqueName: \"kubernetes.io/projected/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-kube-api-access-h8gf7\") pod \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.454972 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-host\") pod \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\" (UID: \"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0\") " Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.455062 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-host" (OuterVolumeSpecName: "host") pod "1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" (UID: "1e4ffb0c-a14c-4204-91f5-af3e3b3832e0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.455457 4829 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-host\") on node \"crc\" DevicePath \"\"" Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.460493 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-kube-api-access-h8gf7" (OuterVolumeSpecName: "kube-api-access-h8gf7") pod "1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" (UID: "1e4ffb0c-a14c-4204-91f5-af3e3b3832e0"). InnerVolumeSpecName "kube-api-access-h8gf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:53:54 crc kubenswrapper[4829]: I1002 08:53:54.557281 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8gf7\" (UniqueName: \"kubernetes.io/projected/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0-kube-api-access-h8gf7\") on node \"crc\" DevicePath \"\"" Oct 02 08:53:55 crc kubenswrapper[4829]: I1002 08:53:55.273081 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" event={"ID":"1e4ffb0c-a14c-4204-91f5-af3e3b3832e0","Type":"ContainerDied","Data":"e59ff5aa9d5c69084cc173e351a931a2673655636396e8b938350e64a57a24fd"} Oct 02 08:53:55 crc kubenswrapper[4829]: I1002 08:53:55.273128 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e59ff5aa9d5c69084cc173e351a931a2673655636396e8b938350e64a57a24fd" Oct 02 08:53:55 crc kubenswrapper[4829]: I1002 08:53:55.273190 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-xv8hz" Oct 02 08:53:58 crc kubenswrapper[4829]: I1002 08:53:58.412749 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:58 crc kubenswrapper[4829]: I1002 08:53:58.464665 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:53:58 crc kubenswrapper[4829]: I1002 08:53:58.644536 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6p926"] Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.321908 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6p926" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="registry-server" containerID="cri-o://c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331" gracePeriod=2 Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.780144 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.822163 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-xv8hz"] Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.829491 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-xv8hz"] Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.864015 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnbts\" (UniqueName: \"kubernetes.io/projected/b4a659f4-409d-434c-b3de-ef23329bf386-kube-api-access-fnbts\") pod \"b4a659f4-409d-434c-b3de-ef23329bf386\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.864394 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-utilities\") pod \"b4a659f4-409d-434c-b3de-ef23329bf386\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.864563 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-catalog-content\") pod \"b4a659f4-409d-434c-b3de-ef23329bf386\" (UID: \"b4a659f4-409d-434c-b3de-ef23329bf386\") " Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.866081 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-utilities" (OuterVolumeSpecName: "utilities") pod "b4a659f4-409d-434c-b3de-ef23329bf386" (UID: "b4a659f4-409d-434c-b3de-ef23329bf386"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.869417 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a659f4-409d-434c-b3de-ef23329bf386-kube-api-access-fnbts" (OuterVolumeSpecName: "kube-api-access-fnbts") pod "b4a659f4-409d-434c-b3de-ef23329bf386" (UID: "b4a659f4-409d-434c-b3de-ef23329bf386"). InnerVolumeSpecName "kube-api-access-fnbts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.946262 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4a659f4-409d-434c-b3de-ef23329bf386" (UID: "b4a659f4-409d-434c-b3de-ef23329bf386"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.966674 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.966885 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnbts\" (UniqueName: \"kubernetes.io/projected/b4a659f4-409d-434c-b3de-ef23329bf386-kube-api-access-fnbts\") on node \"crc\" DevicePath \"\"" Oct 02 08:54:00 crc kubenswrapper[4829]: I1002 08:54:00.966962 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a659f4-409d-434c-b3de-ef23329bf386-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.339679 4829 generic.go:334] "Generic (PLEG): container finished" podID="b4a659f4-409d-434c-b3de-ef23329bf386" containerID="c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331" exitCode=0 Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.339817 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6p926" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.340442 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerDied","Data":"c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331"} Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.340515 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6p926" event={"ID":"b4a659f4-409d-434c-b3de-ef23329bf386","Type":"ContainerDied","Data":"d153a449000ba99879c77dcba4bfc1ebabe64ad0f43c508828916102393377c4"} Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.340541 4829 scope.go:117] "RemoveContainer" containerID="c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.341311 4829 scope.go:117] "RemoveContainer" containerID="d1db2b893a28c7d894fc4fd7db90b2bd511700b6eba9b24fa8ca6def0e4a3321" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.375954 4829 scope.go:117] "RemoveContainer" containerID="38543c24db83a4b3ee983dce25838c71cf0fc649c8a98aff0b14bbdef8feaa87" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.391403 4829 scope.go:117] "RemoveContainer" containerID="ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.428682 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6p926"] Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.429053 4829 scope.go:117] "RemoveContainer" containerID="cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.437785 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6p926"] Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.441622 4829 scope.go:117] "RemoveContainer" containerID="2c1d7f541aee8a8a646115baa4f41309b5d3426c46534d44d6105e18157c331e" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.489703 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" path="/var/lib/kubelet/pods/1e4ffb0c-a14c-4204-91f5-af3e3b3832e0/volumes" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.490244 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" path="/var/lib/kubelet/pods/b4a659f4-409d-434c-b3de-ef23329bf386/volumes" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.503474 4829 scope.go:117] "RemoveContainer" containerID="c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331" Oct 02 08:54:01 crc kubenswrapper[4829]: E1002 08:54:01.503987 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331\": container with ID starting with c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331 not found: ID does not exist" containerID="c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.504052 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331"} err="failed to get container status \"c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331\": rpc error: code = NotFound desc = could not find container \"c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331\": container with ID starting with c8a9c7aea922b2071adc31783601d388bf52bdef19ca466811e72ccfc214b331 not found: ID does not exist" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.504088 4829 scope.go:117] "RemoveContainer" containerID="ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54" Oct 02 08:54:01 crc kubenswrapper[4829]: E1002 08:54:01.504986 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54\": container with ID starting with ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54 not found: ID does not exist" containerID="ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.505027 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54"} err="failed to get container status \"ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54\": rpc error: code = NotFound desc = could not find container \"ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54\": container with ID starting with ddc366e34ba89804ddcbf1dbdcccc503a174bc3da2a21d98cf3588360281ef54 not found: ID does not exist" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.505054 4829 scope.go:117] "RemoveContainer" containerID="cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2" Oct 02 08:54:01 crc kubenswrapper[4829]: E1002 08:54:01.505381 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2\": container with ID starting with cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2 not found: ID does not exist" containerID="cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2" Oct 02 08:54:01 crc kubenswrapper[4829]: I1002 08:54:01.505410 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2"} err="failed to get container status \"cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2\": rpc error: code = NotFound desc = could not find container \"cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2\": container with ID starting with cf711236150c2774fab9b5d4c7dadd90daf2aa72ffc254b5cf277a57854a2bf2 not found: ID does not exist" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.037309 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-lvrkj"] Oct 02 08:54:02 crc kubenswrapper[4829]: E1002 08:54:02.038362 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="registry-server" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.038396 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="registry-server" Oct 02 08:54:02 crc kubenswrapper[4829]: E1002 08:54:02.038456 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" containerName="container-00" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.038469 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" containerName="container-00" Oct 02 08:54:02 crc kubenswrapper[4829]: E1002 08:54:02.038507 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="extract-utilities" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.038520 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="extract-utilities" Oct 02 08:54:02 crc kubenswrapper[4829]: E1002 08:54:02.038546 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="extract-content" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.038558 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="extract-content" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.038939 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e4ffb0c-a14c-4204-91f5-af3e3b3832e0" containerName="container-00" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.038973 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a659f4-409d-434c-b3de-ef23329bf386" containerName="registry-server" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.040111 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.203371 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2srm\" (UniqueName: \"kubernetes.io/projected/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-kube-api-access-w2srm\") pod \"crc-debug-lvrkj\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.203600 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-host\") pod \"crc-debug-lvrkj\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.306167 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-host\") pod \"crc-debug-lvrkj\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.306416 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-host\") pod \"crc-debug-lvrkj\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.306604 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2srm\" (UniqueName: \"kubernetes.io/projected/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-kube-api-access-w2srm\") pod \"crc-debug-lvrkj\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.348415 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2srm\" (UniqueName: \"kubernetes.io/projected/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-kube-api-access-w2srm\") pod \"crc-debug-lvrkj\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: I1002 08:54:02.367623 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:02 crc kubenswrapper[4829]: W1002 08:54:02.414562 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod576bc6f6_dfb1_444d_9ecc_d1c8c0a72a8a.slice/crio-9b8b70a9f194d3466f40c2e7e84b96f467392660b8d54c56b02e4d2298b7ba0a WatchSource:0}: Error finding container 9b8b70a9f194d3466f40c2e7e84b96f467392660b8d54c56b02e4d2298b7ba0a: Status 404 returned error can't find the container with id 9b8b70a9f194d3466f40c2e7e84b96f467392660b8d54c56b02e4d2298b7ba0a Oct 02 08:54:03 crc kubenswrapper[4829]: I1002 08:54:03.370888 4829 generic.go:334] "Generic (PLEG): container finished" podID="576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" containerID="20db13cb8bbe05ca8adbb521b1aa839e15ef43a91d1ef8ae2e137f7fa4259613" exitCode=0 Oct 02 08:54:03 crc kubenswrapper[4829]: I1002 08:54:03.370959 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" event={"ID":"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a","Type":"ContainerDied","Data":"20db13cb8bbe05ca8adbb521b1aa839e15ef43a91d1ef8ae2e137f7fa4259613"} Oct 02 08:54:03 crc kubenswrapper[4829]: I1002 08:54:03.371272 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" event={"ID":"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a","Type":"ContainerStarted","Data":"9b8b70a9f194d3466f40c2e7e84b96f467392660b8d54c56b02e4d2298b7ba0a"} Oct 02 08:54:03 crc kubenswrapper[4829]: I1002 08:54:03.433014 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-lvrkj"] Oct 02 08:54:03 crc kubenswrapper[4829]: I1002 08:54:03.443556 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5t9l/crc-debug-lvrkj"] Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.474095 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.554889 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-host\") pod \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.554975 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-host" (OuterVolumeSpecName: "host") pod "576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" (UID: "576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.555116 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2srm\" (UniqueName: \"kubernetes.io/projected/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-kube-api-access-w2srm\") pod \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\" (UID: \"576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a\") " Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.555977 4829 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-host\") on node \"crc\" DevicePath \"\"" Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.560272 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-kube-api-access-w2srm" (OuterVolumeSpecName: "kube-api-access-w2srm") pod "576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" (UID: "576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a"). InnerVolumeSpecName "kube-api-access-w2srm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:54:04 crc kubenswrapper[4829]: I1002 08:54:04.658143 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2srm\" (UniqueName: \"kubernetes.io/projected/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a-kube-api-access-w2srm\") on node \"crc\" DevicePath \"\"" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.078067 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/util/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.259713 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/pull/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.273867 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/util/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.288577 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/pull/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.387985 4829 scope.go:117] "RemoveContainer" containerID="20db13cb8bbe05ca8adbb521b1aa839e15ef43a91d1ef8ae2e137f7fa4259613" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.388029 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/crc-debug-lvrkj" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.460885 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:54:05 crc kubenswrapper[4829]: E1002 08:54:05.461218 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.471052 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" path="/var/lib/kubelet/pods/576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a/volumes" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.472175 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/pull/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.475101 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/util/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.485691 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/extract/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.660357 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-nf7gj_895426fb-40c5-426b-ab48-f5ac31d22e06/kube-rbac-proxy/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.693442 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-nf7gj_895426fb-40c5-426b-ab48-f5ac31d22e06/manager/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.720210 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-7wrzt_f588d058-d57b-4824-b59d-40731650907f/kube-rbac-proxy/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.870299 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-7wrzt_f588d058-d57b-4824-b59d-40731650907f/manager/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.899461 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-rs8gs_349c96b5-e59f-4eac-8646-e5459775391c/kube-rbac-proxy/0.log" Oct 02 08:54:05 crc kubenswrapper[4829]: I1002 08:54:05.902089 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-rs8gs_349c96b5-e59f-4eac-8646-e5459775391c/manager/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.087065 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-5r5zb_104543c5-16f7-47d1-b2cb-dedd729a87b7/kube-rbac-proxy/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.197022 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-5r5zb_104543c5-16f7-47d1-b2cb-dedd729a87b7/manager/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.275004 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-ztngc_6f685eab-a458-4058-9453-5c4a389758e0/kube-rbac-proxy/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.311135 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-ztngc_6f685eab-a458-4058-9453-5c4a389758e0/manager/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.387782 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-cw4k8_5607af74-7a80-4b1c-9b33-690d3f7fe974/kube-rbac-proxy/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.461491 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-cw4k8_5607af74-7a80-4b1c-9b33-690d3f7fe974/manager/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.615601 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-zc9kp_e46a78e2-3b2a-47c5-8a9a-03877339340d/kube-rbac-proxy/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.721063 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-zc9kp_e46a78e2-3b2a-47c5-8a9a-03877339340d/manager/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.732814 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-2njwr_3f4b21a4-b546-4f69-a432-05ba29b24085/kube-rbac-proxy/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.804590 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-2njwr_3f4b21a4-b546-4f69-a432-05ba29b24085/manager/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.898823 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pf5r7_88cfec3f-d6c0-4007-a845-a3c4e56bcb63/kube-rbac-proxy/0.log" Oct 02 08:54:06 crc kubenswrapper[4829]: I1002 08:54:06.963193 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pf5r7_88cfec3f-d6c0-4007-a845-a3c4e56bcb63/manager/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.031791 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-tmszn_c555f81b-a5f9-4b4b-9ae0-5253ebb12099/kube-rbac-proxy/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.076346 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-tmszn_c555f81b-a5f9-4b4b-9ae0-5253ebb12099/manager/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.248154 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-p8wc8_3430c81b-20da-4871-9f8c-a408573677ec/manager/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.262732 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-p8wc8_3430c81b-20da-4871-9f8c-a408573677ec/kube-rbac-proxy/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.358177 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-2rk2r_d697f00d-25f4-48fb-808e-1a5fb90ff5fa/kube-rbac-proxy/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.471265 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-2rk2r_d697f00d-25f4-48fb-808e-1a5fb90ff5fa/manager/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.515981 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-722dn_bf88c28f-c14a-4390-83ed-75dc3e41061d/kube-rbac-proxy/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.661153 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-722dn_bf88c28f-c14a-4390-83ed-75dc3e41061d/manager/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.713331 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-hlq5c_c35efaf9-0b1b-4801-81c9-f8db24083049/kube-rbac-proxy/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.770653 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-hlq5c_c35efaf9-0b1b-4801-81c9-f8db24083049/manager/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.857254 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-swsvz_2ba71fca-3e78-49e5-bfe5-b831568118ba/kube-rbac-proxy/0.log" Oct 02 08:54:07 crc kubenswrapper[4829]: I1002 08:54:07.899180 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-swsvz_2ba71fca-3e78-49e5-bfe5-b831568118ba/manager/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.025988 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f874976-r72h2_f3a3eba9-d7d0-4796-acdc-2b9525a17e7a/kube-rbac-proxy/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.173487 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6f46d87668-x4x8h_507314d1-5e67-4900-99bd-fd445866f35b/kube-rbac-proxy/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.395253 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6f46d87668-x4x8h_507314d1-5e67-4900-99bd-fd445866f35b/operator/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.400336 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-bqsd6_fba8d39c-ec14-48e6-a9de-41f33e873f09/registry-server/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.612614 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-5cqsw_3dee869a-23fe-4926-a561-67a87ca83103/kube-rbac-proxy/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.679015 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-5cqsw_3dee869a-23fe-4926-a561-67a87ca83103/manager/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.822990 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-vv8cw_8dd30f73-72f3-4733-b5c8-f96753750fb8/kube-rbac-proxy/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.858058 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-vv8cw_8dd30f73-72f3-4733-b5c8-f96753750fb8/manager/0.log" Oct 02 08:54:08 crc kubenswrapper[4829]: I1002 08:54:08.933299 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-d4t94_1c017022-5036-4312-b914-8aeb5439a9ef/operator/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.133807 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-mqflt_fb49bfb5-3c2c-45eb-94d5-8061c919435d/kube-rbac-proxy/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.147296 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-mqflt_fb49bfb5-3c2c-45eb-94d5-8061c919435d/manager/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.198823 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f874976-r72h2_f3a3eba9-d7d0-4796-acdc-2b9525a17e7a/manager/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.314410 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-4wdcm_71170543-3bd9-4d9f-9ad4-a5978a6f018b/kube-rbac-proxy/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.395863 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-twxn6_8129c7b9-335b-4bda-8516-94a818a57591/kube-rbac-proxy/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.440286 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-twxn6_8129c7b9-335b-4bda-8516-94a818a57591/manager/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.553455 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6bf8759486-hrl49_a3ae11b8-1d62-41ba-a63b-4441d2f70709/kube-rbac-proxy/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.560856 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-4wdcm_71170543-3bd9-4d9f-9ad4-a5978a6f018b/manager/0.log" Oct 02 08:54:09 crc kubenswrapper[4829]: I1002 08:54:09.673010 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6bf8759486-hrl49_a3ae11b8-1d62-41ba-a63b-4441d2f70709/manager/0.log" Oct 02 08:54:18 crc kubenswrapper[4829]: I1002 08:54:18.461496 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:54:18 crc kubenswrapper[4829]: E1002 08:54:18.462104 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:54:26 crc kubenswrapper[4829]: I1002 08:54:26.374985 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zqd5d_2cc39185-906e-4b18-a359-b1fd011b74f4/control-plane-machine-set-operator/0.log" Oct 02 08:54:26 crc kubenswrapper[4829]: I1002 08:54:26.509118 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hwmtz_6057de60-fc56-49b5-8843-a9e838186747/kube-rbac-proxy/0.log" Oct 02 08:54:26 crc kubenswrapper[4829]: I1002 08:54:26.513441 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hwmtz_6057de60-fc56-49b5-8843-a9e838186747/machine-api-operator/0.log" Oct 02 08:54:29 crc kubenswrapper[4829]: I1002 08:54:29.469013 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:54:29 crc kubenswrapper[4829]: E1002 08:54:29.469704 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:54:39 crc kubenswrapper[4829]: I1002 08:54:39.203627 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-d6q7t_482724e1-76a1-40b1-b5e2-1a51cade2778/cert-manager-controller/0.log" Oct 02 08:54:39 crc kubenswrapper[4829]: I1002 08:54:39.365997 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-tc7zk_13d4846f-c8e7-40a4-b312-3b3236539721/cert-manager-cainjector/0.log" Oct 02 08:54:39 crc kubenswrapper[4829]: I1002 08:54:39.405753 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-f6xpf_adcd7359-41d5-433b-a4d1-aa964b09f997/cert-manager-webhook/0.log" Oct 02 08:54:41 crc kubenswrapper[4829]: I1002 08:54:41.466020 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:54:41 crc kubenswrapper[4829]: E1002 08:54:41.466557 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.023953 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-t5rvr_d9eea7b1-65ab-4699-ad35-15aad5a523f9/nmstate-console-plugin/0.log" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.132435 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kgg6f_6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e/nmstate-handler/0.log" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.205753 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8rhft_3795be47-c66b-4804-b130-33a73c5e57ce/nmstate-metrics/0.log" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.229114 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8rhft_3795be47-c66b-4804-b130-33a73c5e57ce/kube-rbac-proxy/0.log" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.362044 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hl4ld_7bc8f7d2-12dc-43a9-8cae-872d39a4defc/nmstate-operator/0.log" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.407156 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-6q26v_7e10a126-c321-4d0b-9757-080946eddbba/nmstate-webhook/0.log" Oct 02 08:54:52 crc kubenswrapper[4829]: I1002 08:54:52.461538 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:54:52 crc kubenswrapper[4829]: E1002 08:54:52.461820 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:55:01 crc kubenswrapper[4829]: I1002 08:55:01.595795 4829 scope.go:117] "RemoveContainer" containerID="d1035b40e2eb1bc4bdb42b3fbecb5f5d54216bb46e4b16624c58e521b5995bdc" Oct 02 08:55:01 crc kubenswrapper[4829]: I1002 08:55:01.623641 4829 scope.go:117] "RemoveContainer" containerID="45b27a8b9304a5e027b81d5c36a316b6dcc0e0dafcf014fd1f3ac5f5861d8f78" Oct 02 08:55:01 crc kubenswrapper[4829]: I1002 08:55:01.649861 4829 scope.go:117] "RemoveContainer" containerID="d8c840176c06946721ef8de53611456e3a81af1951ccd80bed88a4d56ace5b6d" Oct 02 08:55:04 crc kubenswrapper[4829]: I1002 08:55:04.461177 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:55:04 crc kubenswrapper[4829]: E1002 08:55:04.462180 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.228295 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-mpn22_321a9e53-9458-4d2f-9784-15f96c7d157b/kube-rbac-proxy/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.458285 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-mpn22_321a9e53-9458-4d2f-9784-15f96c7d157b/controller/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.459825 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.668648 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.685771 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.705169 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.725070 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.924022 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.959053 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.963590 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 08:55:06 crc kubenswrapper[4829]: I1002 08:55:06.969498 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.145883 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.174104 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.182010 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.211044 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/controller/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.348596 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/frr-metrics/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.384984 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/kube-rbac-proxy/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.404932 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/kube-rbac-proxy-frr/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.548847 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/reloader/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.736455 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-t78d2_99893ca5-1605-4a55-9ce8-6711e86fa31b/frr-k8s-webhook-server/0.log" Oct 02 08:55:07 crc kubenswrapper[4829]: I1002 08:55:07.890174 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-75cb48855d-jh2hq_65093a97-cc2b-4983-a106-60b3af7783ca/manager/0.log" Oct 02 08:55:08 crc kubenswrapper[4829]: I1002 08:55:08.048659 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6884746998-tvx77_67cefc41-a1b6-4c77-9770-ea2c4f0fae83/webhook-server/0.log" Oct 02 08:55:08 crc kubenswrapper[4829]: I1002 08:55:08.176766 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mvlxd_3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64/kube-rbac-proxy/0.log" Oct 02 08:55:08 crc kubenswrapper[4829]: I1002 08:55:08.680693 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mvlxd_3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64/speaker/0.log" Oct 02 08:55:08 crc kubenswrapper[4829]: I1002 08:55:08.880865 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/frr/0.log" Oct 02 08:55:17 crc kubenswrapper[4829]: I1002 08:55:17.462391 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:55:17 crc kubenswrapper[4829]: E1002 08:55:17.463141 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.390347 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/util/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.576831 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/util/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.581896 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/pull/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.606850 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/pull/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.749264 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/extract/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.758613 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/util/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.777359 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/pull/0.log" Oct 02 08:55:22 crc kubenswrapper[4829]: I1002 08:55:22.935498 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/util/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.116381 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/pull/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.126400 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/util/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.175644 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/pull/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.315460 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/util/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.323016 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/extract/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.353954 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/pull/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.486597 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-utilities/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.662656 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-content/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.677585 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-content/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.677721 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-utilities/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.828625 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-content/0.log" Oct 02 08:55:23 crc kubenswrapper[4829]: I1002 08:55:23.857032 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-utilities/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.274578 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-utilities/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.436062 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-content/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.486766 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-content/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.486991 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-utilities/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.630741 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/registry-server/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.715140 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-utilities/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.763286 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-content/0.log" Oct 02 08:55:24 crc kubenswrapper[4829]: I1002 08:55:24.908105 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/util/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.220310 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/util/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.259678 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/pull/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.270194 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/pull/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.447866 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/pull/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.455341 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/util/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.479558 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/registry-server/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.502491 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/extract/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.607312 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bbdf8_156b7e29-6cb2-4979-9ef3-c9f7904ecc0f/marketplace-operator/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.718765 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-utilities/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.870121 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-content/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.880840 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-utilities/0.log" Oct 02 08:55:25 crc kubenswrapper[4829]: I1002 08:55:25.905025 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-content/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.069795 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-utilities/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.085815 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-content/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.158782 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-utilities/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.213412 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/registry-server/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.301106 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-utilities/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.349333 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-content/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.351577 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-content/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.494302 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-content/0.log" Oct 02 08:55:26 crc kubenswrapper[4829]: I1002 08:55:26.516696 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-utilities/0.log" Oct 02 08:55:27 crc kubenswrapper[4829]: I1002 08:55:27.209293 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/registry-server/0.log" Oct 02 08:55:32 crc kubenswrapper[4829]: I1002 08:55:32.461180 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:55:32 crc kubenswrapper[4829]: E1002 08:55:32.461969 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:55:39 crc kubenswrapper[4829]: I1002 08:55:39.493187 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-448hn_873d1771-f008-4ac1-8ece-e3b48ba37a81/prometheus-operator/0.log" Oct 02 08:55:39 crc kubenswrapper[4829]: I1002 08:55:39.663965 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv_9f5f2c2d-e40c-4b14-97df-a5a83489726a/prometheus-operator-admission-webhook/0.log" Oct 02 08:55:39 crc kubenswrapper[4829]: I1002 08:55:39.687361 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz_c1ddcd62-beba-4339-9aa1-df07ce7577a8/prometheus-operator-admission-webhook/0.log" Oct 02 08:55:39 crc kubenswrapper[4829]: I1002 08:55:39.857754 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-mw7fx_eb1be399-f44c-4acb-81ad-9723071d7f9f/perses-operator/0.log" Oct 02 08:55:39 crc kubenswrapper[4829]: I1002 08:55:39.879981 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-6th2l_9175ba03-d3bd-4278-8943-1499754f510b/operator/0.log" Oct 02 08:55:45 crc kubenswrapper[4829]: I1002 08:55:45.468159 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:55:45 crc kubenswrapper[4829]: E1002 08:55:45.470039 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 08:55:59 crc kubenswrapper[4829]: I1002 08:55:59.472158 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:56:00 crc kubenswrapper[4829]: I1002 08:56:00.540704 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"64097f1b48a317b51134dfe4c97d1b7feba293cedd22314bc31de6b95e27b269"} Oct 02 08:57:42 crc kubenswrapper[4829]: I1002 08:57:42.864873 4829 generic.go:334] "Generic (PLEG): container finished" podID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerID="3e6112ec904a756317da0aeec6c284d41482e67647573e20b7f40f333aea4534" exitCode=0 Oct 02 08:57:42 crc kubenswrapper[4829]: I1002 08:57:42.864942 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" event={"ID":"b1c62e6e-1e8e-4015-8b54-4179909c1c9b","Type":"ContainerDied","Data":"3e6112ec904a756317da0aeec6c284d41482e67647573e20b7f40f333aea4534"} Oct 02 08:57:42 crc kubenswrapper[4829]: I1002 08:57:42.866507 4829 scope.go:117] "RemoveContainer" containerID="3e6112ec904a756317da0aeec6c284d41482e67647573e20b7f40f333aea4534" Oct 02 08:57:43 crc kubenswrapper[4829]: I1002 08:57:43.819415 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5t9l_must-gather-vbb8f_b1c62e6e-1e8e-4015-8b54-4179909c1c9b/gather/0.log" Oct 02 08:57:52 crc kubenswrapper[4829]: I1002 08:57:52.758477 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h5t9l/must-gather-vbb8f"] Oct 02 08:57:52 crc kubenswrapper[4829]: I1002 08:57:52.760134 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="copy" containerID="cri-o://22dcb65ed5b103d53342b86445aca363fde7b72218ba220c97ac62c5fb360787" gracePeriod=2 Oct 02 08:57:52 crc kubenswrapper[4829]: I1002 08:57:52.767380 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h5t9l/must-gather-vbb8f"] Oct 02 08:57:52 crc kubenswrapper[4829]: I1002 08:57:52.983649 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5t9l_must-gather-vbb8f_b1c62e6e-1e8e-4015-8b54-4179909c1c9b/copy/0.log" Oct 02 08:57:52 crc kubenswrapper[4829]: I1002 08:57:52.984436 4829 generic.go:334] "Generic (PLEG): container finished" podID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerID="22dcb65ed5b103d53342b86445aca363fde7b72218ba220c97ac62c5fb360787" exitCode=143 Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.208370 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5t9l_must-gather-vbb8f_b1c62e6e-1e8e-4015-8b54-4179909c1c9b/copy/0.log" Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.208885 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.314444 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snd7c\" (UniqueName: \"kubernetes.io/projected/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-kube-api-access-snd7c\") pod \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.314599 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-must-gather-output\") pod \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\" (UID: \"b1c62e6e-1e8e-4015-8b54-4179909c1c9b\") " Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.321477 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-kube-api-access-snd7c" (OuterVolumeSpecName: "kube-api-access-snd7c") pod "b1c62e6e-1e8e-4015-8b54-4179909c1c9b" (UID: "b1c62e6e-1e8e-4015-8b54-4179909c1c9b"). InnerVolumeSpecName "kube-api-access-snd7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.416929 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snd7c\" (UniqueName: \"kubernetes.io/projected/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-kube-api-access-snd7c\") on node \"crc\" DevicePath \"\"" Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.557656 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "b1c62e6e-1e8e-4015-8b54-4179909c1c9b" (UID: "b1c62e6e-1e8e-4015-8b54-4179909c1c9b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:57:53 crc kubenswrapper[4829]: I1002 08:57:53.623604 4829 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b1c62e6e-1e8e-4015-8b54-4179909c1c9b-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 08:57:54 crc kubenswrapper[4829]: I1002 08:57:54.009540 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h5t9l_must-gather-vbb8f_b1c62e6e-1e8e-4015-8b54-4179909c1c9b/copy/0.log" Oct 02 08:57:54 crc kubenswrapper[4829]: I1002 08:57:54.010072 4829 scope.go:117] "RemoveContainer" containerID="22dcb65ed5b103d53342b86445aca363fde7b72218ba220c97ac62c5fb360787" Oct 02 08:57:54 crc kubenswrapper[4829]: I1002 08:57:54.010170 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h5t9l/must-gather-vbb8f" Oct 02 08:57:54 crc kubenswrapper[4829]: I1002 08:57:54.073041 4829 scope.go:117] "RemoveContainer" containerID="3e6112ec904a756317da0aeec6c284d41482e67647573e20b7f40f333aea4534" Oct 02 08:57:55 crc kubenswrapper[4829]: I1002 08:57:55.482924 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" path="/var/lib/kubelet/pods/b1c62e6e-1e8e-4015-8b54-4179909c1c9b/volumes" Oct 02 08:58:25 crc kubenswrapper[4829]: I1002 08:58:25.328769 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:58:25 crc kubenswrapper[4829]: I1002 08:58:25.329397 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.800323 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j2qxk"] Oct 02 08:58:35 crc kubenswrapper[4829]: E1002 08:58:35.801310 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="gather" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.801325 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="gather" Oct 02 08:58:35 crc kubenswrapper[4829]: E1002 08:58:35.801351 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="copy" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.801357 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="copy" Oct 02 08:58:35 crc kubenswrapper[4829]: E1002 08:58:35.801374 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" containerName="container-00" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.801380 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" containerName="container-00" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.801563 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="576bc6f6-dfb1-444d-9ecc-d1c8c0a72a8a" containerName="container-00" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.801583 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="gather" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.801596 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1c62e6e-1e8e-4015-8b54-4179909c1c9b" containerName="copy" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.802956 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.816984 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2qxk"] Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.971335 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz2ls\" (UniqueName: \"kubernetes.io/projected/17091d99-a5f9-4c35-baca-526fb4ea550c-kube-api-access-xz2ls\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.971651 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-utilities\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:35 crc kubenswrapper[4829]: I1002 08:58:35.971757 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-catalog-content\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.074077 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz2ls\" (UniqueName: \"kubernetes.io/projected/17091d99-a5f9-4c35-baca-526fb4ea550c-kube-api-access-xz2ls\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.074130 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-utilities\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.074214 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-catalog-content\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.074976 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-catalog-content\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.075685 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-utilities\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.097119 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz2ls\" (UniqueName: \"kubernetes.io/projected/17091d99-a5f9-4c35-baca-526fb4ea550c-kube-api-access-xz2ls\") pod \"redhat-marketplace-j2qxk\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.130114 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:36 crc kubenswrapper[4829]: W1002 08:58:36.604892 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e WatchSource:0}: Error finding container 8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e: Status 404 returned error can't find the container with id 8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e Oct 02 08:58:36 crc kubenswrapper[4829]: I1002 08:58:36.615359 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2qxk"] Oct 02 08:58:37 crc kubenswrapper[4829]: I1002 08:58:37.550103 4829 generic.go:334] "Generic (PLEG): container finished" podID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerID="e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd" exitCode=0 Oct 02 08:58:37 crc kubenswrapper[4829]: I1002 08:58:37.550315 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerDied","Data":"e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd"} Oct 02 08:58:37 crc kubenswrapper[4829]: I1002 08:58:37.550649 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerStarted","Data":"8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e"} Oct 02 08:58:38 crc kubenswrapper[4829]: I1002 08:58:38.568078 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerStarted","Data":"73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb"} Oct 02 08:58:39 crc kubenswrapper[4829]: I1002 08:58:39.586270 4829 generic.go:334] "Generic (PLEG): container finished" podID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerID="73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb" exitCode=0 Oct 02 08:58:39 crc kubenswrapper[4829]: I1002 08:58:39.586408 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerDied","Data":"73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb"} Oct 02 08:58:39 crc kubenswrapper[4829]: I1002 08:58:39.590624 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 08:58:41 crc kubenswrapper[4829]: I1002 08:58:41.629192 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerStarted","Data":"8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e"} Oct 02 08:58:41 crc kubenswrapper[4829]: I1002 08:58:41.658515 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j2qxk" podStartSLOduration=3.832054266 podStartE2EDuration="6.658494094s" podCreationTimestamp="2025-10-02 08:58:35 +0000 UTC" firstStartedPulling="2025-10-02 08:58:37.553034009 +0000 UTC m=+6108.892682444" lastFinishedPulling="2025-10-02 08:58:40.379473827 +0000 UTC m=+6111.719122272" observedRunningTime="2025-10-02 08:58:41.649025209 +0000 UTC m=+6112.988673624" watchObservedRunningTime="2025-10-02 08:58:41.658494094 +0000 UTC m=+6112.998142509" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.350181 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qnkmf/must-gather-9nssl"] Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.352186 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.356675 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qnkmf"/"openshift-service-ca.crt" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.356898 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-qnkmf"/"kube-root-ca.crt" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.374603 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qnkmf/must-gather-9nssl"] Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.544310 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spcnh\" (UniqueName: \"kubernetes.io/projected/10a628e6-642f-4f9c-9234-809aecaa94de-kube-api-access-spcnh\") pod \"must-gather-9nssl\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.544422 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10a628e6-642f-4f9c-9234-809aecaa94de-must-gather-output\") pod \"must-gather-9nssl\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.645904 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spcnh\" (UniqueName: \"kubernetes.io/projected/10a628e6-642f-4f9c-9234-809aecaa94de-kube-api-access-spcnh\") pod \"must-gather-9nssl\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.646009 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10a628e6-642f-4f9c-9234-809aecaa94de-must-gather-output\") pod \"must-gather-9nssl\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.646570 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10a628e6-642f-4f9c-9234-809aecaa94de-must-gather-output\") pod \"must-gather-9nssl\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.673984 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spcnh\" (UniqueName: \"kubernetes.io/projected/10a628e6-642f-4f9c-9234-809aecaa94de-kube-api-access-spcnh\") pod \"must-gather-9nssl\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:42 crc kubenswrapper[4829]: I1002 08:58:42.678827 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 08:58:43 crc kubenswrapper[4829]: I1002 08:58:43.043878 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-qnkmf/must-gather-9nssl"] Oct 02 08:58:43 crc kubenswrapper[4829]: W1002 08:58:43.052976 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a628e6_642f_4f9c_9234_809aecaa94de.slice/crio-5158c635da81b46a25a9966da042e18e0857a0b01baf972732ff292653b579f9 WatchSource:0}: Error finding container 5158c635da81b46a25a9966da042e18e0857a0b01baf972732ff292653b579f9: Status 404 returned error can't find the container with id 5158c635da81b46a25a9966da042e18e0857a0b01baf972732ff292653b579f9 Oct 02 08:58:43 crc kubenswrapper[4829]: I1002 08:58:43.649954 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/must-gather-9nssl" event={"ID":"10a628e6-642f-4f9c-9234-809aecaa94de","Type":"ContainerStarted","Data":"2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1"} Oct 02 08:58:43 crc kubenswrapper[4829]: I1002 08:58:43.650395 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/must-gather-9nssl" event={"ID":"10a628e6-642f-4f9c-9234-809aecaa94de","Type":"ContainerStarted","Data":"8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336"} Oct 02 08:58:43 crc kubenswrapper[4829]: I1002 08:58:43.650425 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/must-gather-9nssl" event={"ID":"10a628e6-642f-4f9c-9234-809aecaa94de","Type":"ContainerStarted","Data":"5158c635da81b46a25a9966da042e18e0857a0b01baf972732ff292653b579f9"} Oct 02 08:58:43 crc kubenswrapper[4829]: I1002 08:58:43.674491 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qnkmf/must-gather-9nssl" podStartSLOduration=1.674465531 podStartE2EDuration="1.674465531s" podCreationTimestamp="2025-10-02 08:58:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:58:43.662953342 +0000 UTC m=+6115.002601747" watchObservedRunningTime="2025-10-02 08:58:43.674465531 +0000 UTC m=+6115.014113956" Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.130789 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.131169 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.185565 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.730555 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.776461 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2qxk"] Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.909308 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-5v2s9"] Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.910823 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:46 crc kubenswrapper[4829]: I1002 08:58:46.912910 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qnkmf"/"default-dockercfg-2nbkl" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.049893 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwz62\" (UniqueName: \"kubernetes.io/projected/0f264c5b-8e55-4d71-9850-0763d9b6ed12-kube-api-access-jwz62\") pod \"crc-debug-5v2s9\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.050349 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f264c5b-8e55-4d71-9850-0763d9b6ed12-host\") pod \"crc-debug-5v2s9\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.152368 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f264c5b-8e55-4d71-9850-0763d9b6ed12-host\") pod \"crc-debug-5v2s9\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.153329 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwz62\" (UniqueName: \"kubernetes.io/projected/0f264c5b-8e55-4d71-9850-0763d9b6ed12-kube-api-access-jwz62\") pod \"crc-debug-5v2s9\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.152480 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f264c5b-8e55-4d71-9850-0763d9b6ed12-host\") pod \"crc-debug-5v2s9\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.179055 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwz62\" (UniqueName: \"kubernetes.io/projected/0f264c5b-8e55-4d71-9850-0763d9b6ed12-kube-api-access-jwz62\") pod \"crc-debug-5v2s9\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.235334 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 08:58:47 crc kubenswrapper[4829]: W1002 08:58:47.275940 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f264c5b_8e55_4d71_9850_0763d9b6ed12.slice/crio-3d2b6ba2265bd64f68e84bba7a8dfb6e0facfce22e740cb52b091b206567bc6f WatchSource:0}: Error finding container 3d2b6ba2265bd64f68e84bba7a8dfb6e0facfce22e740cb52b091b206567bc6f: Status 404 returned error can't find the container with id 3d2b6ba2265bd64f68e84bba7a8dfb6e0facfce22e740cb52b091b206567bc6f Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.699281 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" event={"ID":"0f264c5b-8e55-4d71-9850-0763d9b6ed12","Type":"ContainerStarted","Data":"44cf2611d099cf1e9e878bb9837af983a1f98791d07fcb7447b24d7cb68296e2"} Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.699596 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" event={"ID":"0f264c5b-8e55-4d71-9850-0763d9b6ed12","Type":"ContainerStarted","Data":"3d2b6ba2265bd64f68e84bba7a8dfb6e0facfce22e740cb52b091b206567bc6f"} Oct 02 08:58:47 crc kubenswrapper[4829]: I1002 08:58:47.724352 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" podStartSLOduration=1.724337754 podStartE2EDuration="1.724337754s" podCreationTimestamp="2025-10-02 08:58:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 08:58:47.720071591 +0000 UTC m=+6119.059720006" watchObservedRunningTime="2025-10-02 08:58:47.724337754 +0000 UTC m=+6119.063986159" Oct 02 08:58:48 crc kubenswrapper[4829]: I1002 08:58:48.708587 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j2qxk" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="registry-server" containerID="cri-o://8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e" gracePeriod=2 Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.193095 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.291451 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-utilities\") pod \"17091d99-a5f9-4c35-baca-526fb4ea550c\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.291666 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-catalog-content\") pod \"17091d99-a5f9-4c35-baca-526fb4ea550c\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.291772 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xz2ls\" (UniqueName: \"kubernetes.io/projected/17091d99-a5f9-4c35-baca-526fb4ea550c-kube-api-access-xz2ls\") pod \"17091d99-a5f9-4c35-baca-526fb4ea550c\" (UID: \"17091d99-a5f9-4c35-baca-526fb4ea550c\") " Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.292549 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-utilities" (OuterVolumeSpecName: "utilities") pod "17091d99-a5f9-4c35-baca-526fb4ea550c" (UID: "17091d99-a5f9-4c35-baca-526fb4ea550c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.297218 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17091d99-a5f9-4c35-baca-526fb4ea550c-kube-api-access-xz2ls" (OuterVolumeSpecName: "kube-api-access-xz2ls") pod "17091d99-a5f9-4c35-baca-526fb4ea550c" (UID: "17091d99-a5f9-4c35-baca-526fb4ea550c"). InnerVolumeSpecName "kube-api-access-xz2ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.304018 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17091d99-a5f9-4c35-baca-526fb4ea550c" (UID: "17091d99-a5f9-4c35-baca-526fb4ea550c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.396335 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xz2ls\" (UniqueName: \"kubernetes.io/projected/17091d99-a5f9-4c35-baca-526fb4ea550c-kube-api-access-xz2ls\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.396372 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.396382 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17091d99-a5f9-4c35-baca-526fb4ea550c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.718287 4829 generic.go:334] "Generic (PLEG): container finished" podID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerID="8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e" exitCode=0 Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.718325 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerDied","Data":"8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e"} Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.718753 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j2qxk" event={"ID":"17091d99-a5f9-4c35-baca-526fb4ea550c","Type":"ContainerDied","Data":"8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e"} Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.718772 4829 scope.go:117] "RemoveContainer" containerID="8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.718373 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j2qxk" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.752638 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2qxk"] Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.753176 4829 scope.go:117] "RemoveContainer" containerID="73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.765300 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j2qxk"] Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.774457 4829 scope.go:117] "RemoveContainer" containerID="e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.822748 4829 scope.go:117] "RemoveContainer" containerID="8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e" Oct 02 08:58:49 crc kubenswrapper[4829]: E1002 08:58:49.823164 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e\": container with ID starting with 8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e not found: ID does not exist" containerID="8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.823196 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e"} err="failed to get container status \"8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e\": rpc error: code = NotFound desc = could not find container \"8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e\": container with ID starting with 8acf39d1290e4b9653536538b0571d71deae44dd15fb20820911625806c2be2e not found: ID does not exist" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.823218 4829 scope.go:117] "RemoveContainer" containerID="73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb" Oct 02 08:58:49 crc kubenswrapper[4829]: E1002 08:58:49.823427 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb\": container with ID starting with 73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb not found: ID does not exist" containerID="73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.823448 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb"} err="failed to get container status \"73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb\": rpc error: code = NotFound desc = could not find container \"73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb\": container with ID starting with 73ccadcc9acfd46a4ab5e0e1e973b1842872e2ddf23316d325438e49ad8e9deb not found: ID does not exist" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.823461 4829 scope.go:117] "RemoveContainer" containerID="e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd" Oct 02 08:58:49 crc kubenswrapper[4829]: E1002 08:58:49.823639 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd\": container with ID starting with e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd not found: ID does not exist" containerID="e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd" Oct 02 08:58:49 crc kubenswrapper[4829]: I1002 08:58:49.823660 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd"} err="failed to get container status \"e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd\": rpc error: code = NotFound desc = could not find container \"e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd\": container with ID starting with e1c230ffb7144a3346ac1c9a23712530cdfef08081523db25a2772f303dfddfd not found: ID does not exist" Oct 02 08:58:51 crc kubenswrapper[4829]: I1002 08:58:51.474460 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" path="/var/lib/kubelet/pods/17091d99-a5f9-4c35-baca-526fb4ea550c/volumes" Oct 02 08:58:55 crc kubenswrapper[4829]: I1002 08:58:55.329050 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:58:55 crc kubenswrapper[4829]: I1002 08:58:55.329521 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:58:57 crc kubenswrapper[4829]: E1002 08:58:57.892455 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e\": RecentStats: unable to find data in memory cache]" Oct 02 08:59:08 crc kubenswrapper[4829]: E1002 08:59:08.177282 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:59:18 crc kubenswrapper[4829]: E1002 08:59:18.464895 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e\": RecentStats: unable to find data in memory cache]" Oct 02 08:59:25 crc kubenswrapper[4829]: I1002 08:59:25.328929 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 08:59:25 crc kubenswrapper[4829]: I1002 08:59:25.329555 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 08:59:25 crc kubenswrapper[4829]: I1002 08:59:25.329610 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 08:59:25 crc kubenswrapper[4829]: I1002 08:59:25.330723 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64097f1b48a317b51134dfe4c97d1b7feba293cedd22314bc31de6b95e27b269"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 08:59:25 crc kubenswrapper[4829]: I1002 08:59:25.330783 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://64097f1b48a317b51134dfe4c97d1b7feba293cedd22314bc31de6b95e27b269" gracePeriod=600 Oct 02 08:59:26 crc kubenswrapper[4829]: I1002 08:59:26.048133 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="64097f1b48a317b51134dfe4c97d1b7feba293cedd22314bc31de6b95e27b269" exitCode=0 Oct 02 08:59:26 crc kubenswrapper[4829]: I1002 08:59:26.048279 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"64097f1b48a317b51134dfe4c97d1b7feba293cedd22314bc31de6b95e27b269"} Oct 02 08:59:26 crc kubenswrapper[4829]: I1002 08:59:26.048724 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerStarted","Data":"638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c"} Oct 02 08:59:26 crc kubenswrapper[4829]: I1002 08:59:26.048754 4829 scope.go:117] "RemoveContainer" containerID="06e61cd8b37985723be04cbb0c3f82d42aad960b43dcacd55f9781541233a592" Oct 02 08:59:28 crc kubenswrapper[4829]: E1002 08:59:28.700175 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:59:38 crc kubenswrapper[4829]: E1002 08:59:38.952993 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice\": RecentStats: unable to find data in memory cache]" Oct 02 08:59:49 crc kubenswrapper[4829]: E1002 08:59:49.217427 4829 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice/crio-8a7dd540ba9a1d62a73f6b0f1c5f9afe882e3f31a1e5cec164e9da33924c716e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17091d99_a5f9_4c35_baca_526fb4ea550c.slice\": RecentStats: unable to find data in memory cache]" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.182266 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv"] Oct 02 09:00:00 crc kubenswrapper[4829]: E1002 09:00:00.183508 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="registry-server" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.183533 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="registry-server" Oct 02 09:00:00 crc kubenswrapper[4829]: E1002 09:00:00.183584 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="extract-utilities" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.183595 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="extract-utilities" Oct 02 09:00:00 crc kubenswrapper[4829]: E1002 09:00:00.183621 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="extract-content" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.183632 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="extract-content" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.183965 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="17091d99-a5f9-4c35-baca-526fb4ea550c" containerName="registry-server" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.185139 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.188450 4829 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.188939 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.201216 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv"] Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.312903 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85dv6\" (UniqueName: \"kubernetes.io/projected/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-kube-api-access-85dv6\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.313090 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-secret-volume\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.313133 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-config-volume\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.415062 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-secret-volume\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.416746 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-config-volume\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.416834 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-config-volume\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.417008 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85dv6\" (UniqueName: \"kubernetes.io/projected/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-kube-api-access-85dv6\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.425139 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-secret-volume\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.455184 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85dv6\" (UniqueName: \"kubernetes.io/projected/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-kube-api-access-85dv6\") pod \"collect-profiles-29323260-96gbv\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:00 crc kubenswrapper[4829]: I1002 09:00:00.517104 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:01 crc kubenswrapper[4829]: I1002 09:00:01.036360 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv"] Oct 02 09:00:01 crc kubenswrapper[4829]: I1002 09:00:01.374592 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" event={"ID":"639ba6ed-adf7-47d7-89bb-8c14ae0d371d","Type":"ContainerStarted","Data":"a9cce5ae3615550cd9e6bc7b2a214bd22d4b40d82bf294e5eff0a4edb4df17e5"} Oct 02 09:00:01 crc kubenswrapper[4829]: I1002 09:00:01.374885 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" event={"ID":"639ba6ed-adf7-47d7-89bb-8c14ae0d371d","Type":"ContainerStarted","Data":"44a677d9e499252f882f24b5ddc116e4ede0b7cd559db7e3232d9afd394a208d"} Oct 02 09:00:01 crc kubenswrapper[4829]: I1002 09:00:01.397462 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" podStartSLOduration=1.397443494 podStartE2EDuration="1.397443494s" podCreationTimestamp="2025-10-02 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:00:01.39375781 +0000 UTC m=+6192.733406225" watchObservedRunningTime="2025-10-02 09:00:01.397443494 +0000 UTC m=+6192.737091909" Oct 02 09:00:01 crc kubenswrapper[4829]: I1002 09:00:01.930346 4829 scope.go:117] "RemoveContainer" containerID="cc3ead9b692ac4dc301248b000e155f1ae4fd243c198c5b4648aa8ede7c60e51" Oct 02 09:00:02 crc kubenswrapper[4829]: I1002 09:00:02.385353 4829 generic.go:334] "Generic (PLEG): container finished" podID="639ba6ed-adf7-47d7-89bb-8c14ae0d371d" containerID="a9cce5ae3615550cd9e6bc7b2a214bd22d4b40d82bf294e5eff0a4edb4df17e5" exitCode=0 Oct 02 09:00:02 crc kubenswrapper[4829]: I1002 09:00:02.385403 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" event={"ID":"639ba6ed-adf7-47d7-89bb-8c14ae0d371d","Type":"ContainerDied","Data":"a9cce5ae3615550cd9e6bc7b2a214bd22d4b40d82bf294e5eff0a4edb4df17e5"} Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.754009 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.830999 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-config-volume\") pod \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.831124 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-secret-volume\") pod \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.831185 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85dv6\" (UniqueName: \"kubernetes.io/projected/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-kube-api-access-85dv6\") pod \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\" (UID: \"639ba6ed-adf7-47d7-89bb-8c14ae0d371d\") " Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.832281 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-config-volume" (OuterVolumeSpecName: "config-volume") pod "639ba6ed-adf7-47d7-89bb-8c14ae0d371d" (UID: "639ba6ed-adf7-47d7-89bb-8c14ae0d371d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.847420 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-kube-api-access-85dv6" (OuterVolumeSpecName: "kube-api-access-85dv6") pod "639ba6ed-adf7-47d7-89bb-8c14ae0d371d" (UID: "639ba6ed-adf7-47d7-89bb-8c14ae0d371d"). InnerVolumeSpecName "kube-api-access-85dv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.852364 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "639ba6ed-adf7-47d7-89bb-8c14ae0d371d" (UID: "639ba6ed-adf7-47d7-89bb-8c14ae0d371d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.934084 4829 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.934130 4829 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:03 crc kubenswrapper[4829]: I1002 09:00:03.934144 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85dv6\" (UniqueName: \"kubernetes.io/projected/639ba6ed-adf7-47d7-89bb-8c14ae0d371d-kube-api-access-85dv6\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:04 crc kubenswrapper[4829]: I1002 09:00:04.404806 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" event={"ID":"639ba6ed-adf7-47d7-89bb-8c14ae0d371d","Type":"ContainerDied","Data":"44a677d9e499252f882f24b5ddc116e4ede0b7cd559db7e3232d9afd394a208d"} Oct 02 09:00:04 crc kubenswrapper[4829]: I1002 09:00:04.405086 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44a677d9e499252f882f24b5ddc116e4ede0b7cd559db7e3232d9afd394a208d" Oct 02 09:00:04 crc kubenswrapper[4829]: I1002 09:00:04.404903 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29323260-96gbv" Oct 02 09:00:04 crc kubenswrapper[4829]: I1002 09:00:04.475755 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj"] Oct 02 09:00:04 crc kubenswrapper[4829]: I1002 09:00:04.492814 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29323215-qh6nj"] Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.249801 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7944ccd6-g5gh7_69f201db-52b1-4a9f-a1f3-adb1bdb923da/barbican-api/0.log" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.291653 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6c7944ccd6-g5gh7_69f201db-52b1-4a9f-a1f3-adb1bdb923da/barbican-api-log/0.log" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.462788 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7cf66b5f98-d625p_1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244/barbican-keystone-listener/0.log" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.471324 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17916ae-ef45-4999-8b6d-42f8476721e9" path="/var/lib/kubelet/pods/d17916ae-ef45-4999-8b6d-42f8476721e9/volumes" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.521300 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7cf66b5f98-d625p_1ebdeb90-eabf-4ef4-8fb4-b8bbfa13f244/barbican-keystone-listener-log/0.log" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.706078 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f9d8b6c5-ds24l_54708879-c00e-4582-a37e-5becf429dc5d/barbican-worker/0.log" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.719299 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f9d8b6c5-ds24l_54708879-c00e-4582-a37e-5becf429dc5d/barbican-worker-log/0.log" Oct 02 09:00:05 crc kubenswrapper[4829]: I1002 09:00:05.895337 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-4hmvk_46cccbf0-9925-4708-bd05-119652ca2732/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.107928 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/ceilometer-central-agent/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.136767 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/ceilometer-notification-agent/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.155432 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/proxy-httpd/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.276108 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_409b81a0-0c77-4550-a88a-8489473fe397/sg-core/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.542913 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2bec1c3c-60b9-4776-93c0-3d51886b3552/cinder-api/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.621776 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_2bec1c3c-60b9-4776-93c0-3d51886b3552/cinder-api-log/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.828785 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5da522c4-7a44-46bc-b84e-eb230aad0bb7/cinder-scheduler/0.log" Oct 02 09:00:06 crc kubenswrapper[4829]: I1002 09:00:06.888240 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5da522c4-7a44-46bc-b84e-eb230aad0bb7/probe/0.log" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.023944 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-j8jwz_e6b2c2b0-9a68-4525-9f18-64b8e4354246/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.225423 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k746m_28b94275-3752-42ab-b862-fee2ee0ead47/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.338375 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8gbjf"] Oct 02 09:00:07 crc kubenswrapper[4829]: E1002 09:00:07.338830 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="639ba6ed-adf7-47d7-89bb-8c14ae0d371d" containerName="collect-profiles" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.338845 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="639ba6ed-adf7-47d7-89bb-8c14ae0d371d" containerName="collect-profiles" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.339122 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="639ba6ed-adf7-47d7-89bb-8c14ae0d371d" containerName="collect-profiles" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.343860 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.356941 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8gbjf"] Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.408395 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-catalog-content\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.408596 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bmzg\" (UniqueName: \"kubernetes.io/projected/50c79898-c3c5-4e82-83c1-b2d7cbf99318-kube-api-access-5bmzg\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.408835 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-utilities\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.408880 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k8v9m_f418c44c-86bc-4ba6-aabb-da19658c0e77/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.510933 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-catalog-content\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.511027 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bmzg\" (UniqueName: \"kubernetes.io/projected/50c79898-c3c5-4e82-83c1-b2d7cbf99318-kube-api-access-5bmzg\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.511111 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-utilities\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.511461 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-catalog-content\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.511508 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-utilities\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.566334 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bmzg\" (UniqueName: \"kubernetes.io/projected/50c79898-c3c5-4e82-83c1-b2d7cbf99318-kube-api-access-5bmzg\") pod \"community-operators-8gbjf\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.668183 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:07 crc kubenswrapper[4829]: I1002 09:00:07.697908 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-cvpg4_0c1b029b-bf2c-4b4b-b71c-6050f640212b/init/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.098459 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-cvpg4_0c1b029b-bf2c-4b4b-b71c-6050f640212b/init/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.221858 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-cvpg4_0c1b029b-bf2c-4b4b-b71c-6050f640212b/dnsmasq-dns/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.290600 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8gbjf"] Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.356996 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-nlhkp_ac409859-eb90-422d-ad2f-4cf60f171844/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.438265 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerStarted","Data":"c466ff2e5e2204e3abdd17ac1f5decc60cd2bffe32ea11a73685446535ff4251"} Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.446988 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3f017e85-e428-48e7-9b00-705ca534019d/glance-httpd/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.473876 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3f017e85-e428-48e7-9b00-705ca534019d/glance-log/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.649336 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_9c3dbee2-4ec4-4d14-8e07-8a04428f0429/glance-httpd/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.692210 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_9c3dbee2-4ec4-4d14-8e07-8a04428f0429/glance-log/0.log" Oct 02 09:00:08 crc kubenswrapper[4829]: I1002 09:00:08.931609 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5df99f8898-76z9s_3407e6aa-fc06-4a2e-bd97-b8540ae12167/horizon/0.log" Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.071142 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-7krkz_28912442-1f56-461e-9d26-18227d1f2386/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.209871 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-dfmsw_4181a86c-75a3-4400-9b89-352c04c06044/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.451628 4829 generic.go:334] "Generic (PLEG): container finished" podID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerID="72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a" exitCode=0 Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.451887 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerDied","Data":"72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a"} Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.663279 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29323201-j6v6q_0ca1b8dd-db65-4bcf-908c-3bf2b145d334/keystone-cron/0.log" Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.719746 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5df99f8898-76z9s_3407e6aa-fc06-4a2e-bd97-b8540ae12167/horizon-log/0.log" Oct 02 09:00:09 crc kubenswrapper[4829]: I1002 09:00:09.896292 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_fadeaa43-1547-4915-9e34-43e880485d84/kube-state-metrics/0.log" Oct 02 09:00:10 crc kubenswrapper[4829]: I1002 09:00:10.099828 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6b77fd5f6d-mmltx_5d8b5f01-0374-4315-82d6-de60c41b6bad/keystone-api/0.log" Oct 02 09:00:10 crc kubenswrapper[4829]: I1002 09:00:10.101052 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-94ftd_9dfa4494-c2fb-415c-a38a-79afbf13be81/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:10 crc kubenswrapper[4829]: I1002 09:00:10.771070 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79b8d665c5-dgn5f_1a40a5fd-a82e-430c-a15d-e32f0a14f5f2/neutron-api/0.log" Oct 02 09:00:10 crc kubenswrapper[4829]: I1002 09:00:10.867823 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-79b8d665c5-dgn5f_1a40a5fd-a82e-430c-a15d-e32f0a14f5f2/neutron-httpd/0.log" Oct 02 09:00:11 crc kubenswrapper[4829]: I1002 09:00:11.016817 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-pq6dc_d0041618-1a4e-474f-8570-ccf383c6cff9/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:11 crc kubenswrapper[4829]: I1002 09:00:11.505693 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerStarted","Data":"39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b"} Oct 02 09:00:11 crc kubenswrapper[4829]: I1002 09:00:11.999432 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_911b568c-7273-4d79-8213-0d2577e98046/nova-cell0-conductor-conductor/0.log" Oct 02 09:00:12 crc kubenswrapper[4829]: I1002 09:00:12.532146 4829 generic.go:334] "Generic (PLEG): container finished" podID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerID="39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b" exitCode=0 Oct 02 09:00:12 crc kubenswrapper[4829]: I1002 09:00:12.532778 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerDied","Data":"39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b"} Oct 02 09:00:12 crc kubenswrapper[4829]: I1002 09:00:12.729509 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23175b6b-8b9a-43da-a7ba-d0b5cb04863a/nova-api-log/0.log" Oct 02 09:00:12 crc kubenswrapper[4829]: I1002 09:00:12.749948 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_483a24aa-87d3-4b0a-accd-5ca4a8a7a029/nova-cell1-conductor-conductor/0.log" Oct 02 09:00:13 crc kubenswrapper[4829]: I1002 09:00:13.301038 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_23a2af5c-929f-41a7-8a96-d42f5902f200/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 09:00:13 crc kubenswrapper[4829]: I1002 09:00:13.413103 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_23175b6b-8b9a-43da-a7ba-d0b5cb04863a/nova-api-api/0.log" Oct 02 09:00:13 crc kubenswrapper[4829]: I1002 09:00:13.543486 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerStarted","Data":"0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896"} Oct 02 09:00:13 crc kubenswrapper[4829]: I1002 09:00:13.571531 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8gbjf" podStartSLOduration=3.021325318 podStartE2EDuration="6.571512186s" podCreationTimestamp="2025-10-02 09:00:07 +0000 UTC" firstStartedPulling="2025-10-02 09:00:09.454140629 +0000 UTC m=+6200.793789034" lastFinishedPulling="2025-10-02 09:00:13.004327497 +0000 UTC m=+6204.343975902" observedRunningTime="2025-10-02 09:00:13.562866906 +0000 UTC m=+6204.902515311" watchObservedRunningTime="2025-10-02 09:00:13.571512186 +0000 UTC m=+6204.911160591" Oct 02 09:00:13 crc kubenswrapper[4829]: I1002 09:00:13.614763 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-jmb99_97915d50-01b1-415e-99f1-17c7d4340c66/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:13 crc kubenswrapper[4829]: I1002 09:00:13.831209 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1298703b-f77d-4cf5-9537-8928493b4e6b/nova-metadata-log/0.log" Oct 02 09:00:14 crc kubenswrapper[4829]: I1002 09:00:14.782199 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d72fc392-aa3b-40de-98cd-a0f7a633d5fa/mysql-bootstrap/0.log" Oct 02 09:00:14 crc kubenswrapper[4829]: I1002 09:00:14.833720 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_52f60623-b76f-43dc-9022-d62914e8950b/nova-scheduler-scheduler/0.log" Oct 02 09:00:15 crc kubenswrapper[4829]: I1002 09:00:15.087411 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d72fc392-aa3b-40de-98cd-a0f7a633d5fa/galera/0.log" Oct 02 09:00:15 crc kubenswrapper[4829]: I1002 09:00:15.090305 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d72fc392-aa3b-40de-98cd-a0f7a633d5fa/mysql-bootstrap/0.log" Oct 02 09:00:15 crc kubenswrapper[4829]: I1002 09:00:15.417652 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0195e48a-f8c9-433b-a9a8-c3da38457280/mysql-bootstrap/0.log" Oct 02 09:00:15 crc kubenswrapper[4829]: I1002 09:00:15.722077 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0195e48a-f8c9-433b-a9a8-c3da38457280/mysql-bootstrap/0.log" Oct 02 09:00:15 crc kubenswrapper[4829]: I1002 09:00:15.790770 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0195e48a-f8c9-433b-a9a8-c3da38457280/galera/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.003947 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_859798c1-faae-46b4-83f2-0b9d18734236/openstackclient/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.214484 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-s5fpj_4d3450e4-72b1-420c-8f0e-3c97a9eaa5f3/openstack-network-exporter/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.439187 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_1298703b-f77d-4cf5-9537-8928493b4e6b/nova-metadata-metadata/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.471139 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-njw55_91ff1e5f-73e2-4237-b201-84065a586553/ovn-controller/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.742764 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovsdb-server-init/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.900906 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovs-vswitchd/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.915022 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovsdb-server/0.log" Oct 02 09:00:16 crc kubenswrapper[4829]: I1002 09:00:16.987973 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-t8kr7_bcd90918-1fff-4cf6-9020-7780c2c53672/ovsdb-server-init/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.179715 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xsv2v_adb0b737-9122-4893-8b4d-2111d18c2a4d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.311946 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_775c00d4-e88e-42a8-9351-edd61e0c2cd6/openstack-network-exporter/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.377056 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_775c00d4-e88e-42a8-9351-edd61e0c2cd6/ovn-northd/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.554835 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_37f64531-608d-4dd5-99ad-5c928bcf7ca0/openstack-network-exporter/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.653809 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_37f64531-608d-4dd5-99ad-5c928bcf7ca0/ovsdbserver-nb/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.668321 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.668373 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.716195 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.783577 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cedf4e9a-cfe0-4499-8600-edf90b9b5cda/openstack-network-exporter/0.log" Oct 02 09:00:17 crc kubenswrapper[4829]: I1002 09:00:17.915971 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cedf4e9a-cfe0-4499-8600-edf90b9b5cda/ovsdbserver-sb/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.431816 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-995ff44cb-bcb2z_f49ffa44-6b66-43b1-9201-afc0e8bb9311/placement-log/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.437053 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-995ff44cb-bcb2z_f49ffa44-6b66-43b1-9201-afc0e8bb9311/placement-api/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.550009 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/init-config-reloader/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.649200 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.706268 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8gbjf"] Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.728056 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/init-config-reloader/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.748777 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/prometheus/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.786245 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/config-reloader/0.log" Oct 02 09:00:18 crc kubenswrapper[4829]: I1002 09:00:18.927799 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_2de0625f-160a-4ca9-865e-35a8f782ed2d/thanos-sidecar/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.001451 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66249bb-0358-455a-8ca1-7c9688cfe5f0/setup-container/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.194791 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66249bb-0358-455a-8ca1-7c9688cfe5f0/setup-container/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.241632 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b66249bb-0358-455a-8ca1-7c9688cfe5f0/rabbitmq/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.393751 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_43522a1c-881d-4d33-91e3-3dea8c241d8c/setup-container/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.645509 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_43522a1c-881d-4d33-91e3-3dea8c241d8c/setup-container/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.651144 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_43522a1c-881d-4d33-91e3-3dea8c241d8c/rabbitmq/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.883371 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-r5t87_4d515775-5135-4f3f-aa21-fd1f648afb11/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:19 crc kubenswrapper[4829]: I1002 09:00:19.935917 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fm777_fb03e152-ef79-4d7f-920d-5d1fbacb7167/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.094216 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-4wl26_01ec7e0b-c62e-4bd4-ab7e-6756c5314161/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.301326 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-2kgpj_3ed1e719-596e-42fd-ac3f-b092996c9ca6/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.382082 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-8b6wr_0506e251-5b72-4ce4-91c5-8116bf387c34/ssh-known-hosts-edpm-deployment/0.log" Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.603162 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b89d7c8c-f2lgb_395b490d-fbc2-403d-ac35-9541329363c2/proxy-server/0.log" Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.616592 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8gbjf" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="registry-server" containerID="cri-o://0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896" gracePeriod=2 Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.808131 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-m6p2d_b5063471-4946-4763-9338-1ad3fbb4b9ef/swift-ring-rebalance/0.log" Oct 02 09:00:20 crc kubenswrapper[4829]: I1002 09:00:20.812292 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-84b89d7c8c-f2lgb_395b490d-fbc2-403d-ac35-9541329363c2/proxy-httpd/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.047519 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-reaper/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.090391 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-auditor/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.129391 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.207010 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bmzg\" (UniqueName: \"kubernetes.io/projected/50c79898-c3c5-4e82-83c1-b2d7cbf99318-kube-api-access-5bmzg\") pod \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.207110 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-utilities\") pod \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.207139 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-catalog-content\") pod \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\" (UID: \"50c79898-c3c5-4e82-83c1-b2d7cbf99318\") " Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.217058 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-utilities" (OuterVolumeSpecName: "utilities") pod "50c79898-c3c5-4e82-83c1-b2d7cbf99318" (UID: "50c79898-c3c5-4e82-83c1-b2d7cbf99318"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.246401 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50c79898-c3c5-4e82-83c1-b2d7cbf99318-kube-api-access-5bmzg" (OuterVolumeSpecName: "kube-api-access-5bmzg") pod "50c79898-c3c5-4e82-83c1-b2d7cbf99318" (UID: "50c79898-c3c5-4e82-83c1-b2d7cbf99318"). InnerVolumeSpecName "kube-api-access-5bmzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.256457 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "50c79898-c3c5-4e82-83c1-b2d7cbf99318" (UID: "50c79898-c3c5-4e82-83c1-b2d7cbf99318"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.310557 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.310587 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50c79898-c3c5-4e82-83c1-b2d7cbf99318-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.310599 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bmzg\" (UniqueName: \"kubernetes.io/projected/50c79898-c3c5-4e82-83c1-b2d7cbf99318-kube-api-access-5bmzg\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.334507 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-server/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.354396 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/account-replicator/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.390577 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-auditor/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.528210 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-server/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.550870 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-replicator/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.583843 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/container-updater/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.628843 4829 generic.go:334] "Generic (PLEG): container finished" podID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerID="0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896" exitCode=0 Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.628880 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gbjf" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.628918 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerDied","Data":"0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896"} Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.629460 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gbjf" event={"ID":"50c79898-c3c5-4e82-83c1-b2d7cbf99318","Type":"ContainerDied","Data":"c466ff2e5e2204e3abdd17ac1f5decc60cd2bffe32ea11a73685446535ff4251"} Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.629492 4829 scope.go:117] "RemoveContainer" containerID="0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.658334 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8gbjf"] Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.664813 4829 scope.go:117] "RemoveContainer" containerID="39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.681724 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8gbjf"] Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.686593 4829 scope.go:117] "RemoveContainer" containerID="72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.744448 4829 scope.go:117] "RemoveContainer" containerID="0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896" Oct 02 09:00:21 crc kubenswrapper[4829]: E1002 09:00:21.747567 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896\": container with ID starting with 0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896 not found: ID does not exist" containerID="0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.747607 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896"} err="failed to get container status \"0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896\": rpc error: code = NotFound desc = could not find container \"0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896\": container with ID starting with 0f47816b4ee18e0d5b43e63dc61672774d710cccf6e95a83635b79cd7f63c896 not found: ID does not exist" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.747633 4829 scope.go:117] "RemoveContainer" containerID="39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b" Oct 02 09:00:21 crc kubenswrapper[4829]: E1002 09:00:21.749120 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b\": container with ID starting with 39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b not found: ID does not exist" containerID="39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.749147 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b"} err="failed to get container status \"39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b\": rpc error: code = NotFound desc = could not find container \"39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b\": container with ID starting with 39e551815c430b7bfa50e4e06eeaba9e4bf15086e8b4e0ab09ddd13c3fae349b not found: ID does not exist" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.749162 4829 scope.go:117] "RemoveContainer" containerID="72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a" Oct 02 09:00:21 crc kubenswrapper[4829]: E1002 09:00:21.749379 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a\": container with ID starting with 72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a not found: ID does not exist" containerID="72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.749418 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a"} err="failed to get container status \"72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a\": rpc error: code = NotFound desc = could not find container \"72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a\": container with ID starting with 72215ef526bb26098e4ecdd6b59262b91ca36cf688f4fb887c750083fbc47f8a not found: ID does not exist" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.787121 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-auditor/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.810125 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-expirer/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.897785 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-replicator/0.log" Oct 02 09:00:21 crc kubenswrapper[4829]: I1002 09:00:21.977344 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-server/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.022362 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/object-updater/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.107748 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/rsync/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.251788 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_9da44bca-624e-49ca-8fd2-e0b0974c9ae5/swift-recon-cron/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.374721 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-s6cdx_e74d0c61-084d-49d1-877a-3325875d5c15/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.474457 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_18e0103e-1a6b-4f9c-b583-f3dc8f405137/tempest-tests-tempest-tests-runner/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.698716 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f6e01654-3f1b-4049-be7e-a562dfaa59df/test-operator-logs-container/0.log" Oct 02 09:00:22 crc kubenswrapper[4829]: I1002 09:00:22.956679 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wrlj5_bb87d295-2bab-44ea-b5cc-cc77a0efe964/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 02 09:00:23 crc kubenswrapper[4829]: I1002 09:00:23.473483 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" path="/var/lib/kubelet/pods/50c79898-c3c5-4e82-83c1-b2d7cbf99318/volumes" Oct 02 09:00:23 crc kubenswrapper[4829]: I1002 09:00:23.871575 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_59d8c82e-9aca-41ca-9d4b-8f5ec2824132/watcher-applier/0.log" Oct 02 09:00:24 crc kubenswrapper[4829]: I1002 09:00:24.212873 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_ab94865b-0ce8-4b10-bbab-7354e3603e1b/watcher-api-log/0.log" Oct 02 09:00:25 crc kubenswrapper[4829]: I1002 09:00:25.410166 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_c6390f00-cfcd-4b1c-831d-fd355daac37e/watcher-decision-engine/0.log" Oct 02 09:00:27 crc kubenswrapper[4829]: I1002 09:00:27.773209 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_ab94865b-0ce8-4b10-bbab-7354e3603e1b/watcher-api/0.log" Oct 02 09:00:30 crc kubenswrapper[4829]: I1002 09:00:30.370785 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5146035b-0eb1-40c3-bc13-9d40affa0f2e/memcached/0.log" Oct 02 09:00:48 crc kubenswrapper[4829]: I1002 09:00:48.915193 4829 generic.go:334] "Generic (PLEG): container finished" podID="0f264c5b-8e55-4d71-9850-0763d9b6ed12" containerID="44cf2611d099cf1e9e878bb9837af983a1f98791d07fcb7447b24d7cb68296e2" exitCode=0 Oct 02 09:00:48 crc kubenswrapper[4829]: I1002 09:00:48.915291 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" event={"ID":"0f264c5b-8e55-4d71-9850-0763d9b6ed12","Type":"ContainerDied","Data":"44cf2611d099cf1e9e878bb9837af983a1f98791d07fcb7447b24d7cb68296e2"} Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.024520 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.055877 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-5v2s9"] Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.064057 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-5v2s9"] Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.139972 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwz62\" (UniqueName: \"kubernetes.io/projected/0f264c5b-8e55-4d71-9850-0763d9b6ed12-kube-api-access-jwz62\") pod \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.140019 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f264c5b-8e55-4d71-9850-0763d9b6ed12-host\") pod \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\" (UID: \"0f264c5b-8e55-4d71-9850-0763d9b6ed12\") " Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.140258 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0f264c5b-8e55-4d71-9850-0763d9b6ed12-host" (OuterVolumeSpecName: "host") pod "0f264c5b-8e55-4d71-9850-0763d9b6ed12" (UID: "0f264c5b-8e55-4d71-9850-0763d9b6ed12"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.140496 4829 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0f264c5b-8e55-4d71-9850-0763d9b6ed12-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.146988 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f264c5b-8e55-4d71-9850-0763d9b6ed12-kube-api-access-jwz62" (OuterVolumeSpecName: "kube-api-access-jwz62") pod "0f264c5b-8e55-4d71-9850-0763d9b6ed12" (UID: "0f264c5b-8e55-4d71-9850-0763d9b6ed12"). InnerVolumeSpecName "kube-api-access-jwz62". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.242575 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwz62\" (UniqueName: \"kubernetes.io/projected/0f264c5b-8e55-4d71-9850-0763d9b6ed12-kube-api-access-jwz62\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.937218 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d2b6ba2265bd64f68e84bba7a8dfb6e0facfce22e740cb52b091b206567bc6f" Oct 02 09:00:50 crc kubenswrapper[4829]: I1002 09:00:50.937286 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-5v2s9" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.420310 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-hmjb7"] Oct 02 09:00:51 crc kubenswrapper[4829]: E1002 09:00:51.420901 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="extract-utilities" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.420914 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="extract-utilities" Oct 02 09:00:51 crc kubenswrapper[4829]: E1002 09:00:51.420935 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="extract-content" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.420940 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="extract-content" Oct 02 09:00:51 crc kubenswrapper[4829]: E1002 09:00:51.420953 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="registry-server" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.420959 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="registry-server" Oct 02 09:00:51 crc kubenswrapper[4829]: E1002 09:00:51.420978 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f264c5b-8e55-4d71-9850-0763d9b6ed12" containerName="container-00" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.420983 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f264c5b-8e55-4d71-9850-0763d9b6ed12" containerName="container-00" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.421153 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f264c5b-8e55-4d71-9850-0763d9b6ed12" containerName="container-00" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.421181 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="50c79898-c3c5-4e82-83c1-b2d7cbf99318" containerName="registry-server" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.421842 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.423678 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qnkmf"/"default-dockercfg-2nbkl" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.474109 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f264c5b-8e55-4d71-9850-0763d9b6ed12" path="/var/lib/kubelet/pods/0f264c5b-8e55-4d71-9850-0763d9b6ed12/volumes" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.565975 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnp2c\" (UniqueName: \"kubernetes.io/projected/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-kube-api-access-fnp2c\") pod \"crc-debug-hmjb7\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.566141 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-host\") pod \"crc-debug-hmjb7\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.667754 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnp2c\" (UniqueName: \"kubernetes.io/projected/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-kube-api-access-fnp2c\") pod \"crc-debug-hmjb7\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.667869 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-host\") pod \"crc-debug-hmjb7\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.668037 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-host\") pod \"crc-debug-hmjb7\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.685882 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnp2c\" (UniqueName: \"kubernetes.io/projected/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-kube-api-access-fnp2c\") pod \"crc-debug-hmjb7\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.741862 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:51 crc kubenswrapper[4829]: W1002 09:00:51.769868 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05cf5f68_7cfc_4caa_8db3_540b2c34caa0.slice/crio-d0b86e61a087599dc5591acbadc8852ed55cf49f2930f4ce19dad8bdaeb24508 WatchSource:0}: Error finding container d0b86e61a087599dc5591acbadc8852ed55cf49f2930f4ce19dad8bdaeb24508: Status 404 returned error can't find the container with id d0b86e61a087599dc5591acbadc8852ed55cf49f2930f4ce19dad8bdaeb24508 Oct 02 09:00:51 crc kubenswrapper[4829]: I1002 09:00:51.953375 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" event={"ID":"05cf5f68-7cfc-4caa-8db3-540b2c34caa0","Type":"ContainerStarted","Data":"d0b86e61a087599dc5591acbadc8852ed55cf49f2930f4ce19dad8bdaeb24508"} Oct 02 09:00:52 crc kubenswrapper[4829]: I1002 09:00:52.968359 4829 generic.go:334] "Generic (PLEG): container finished" podID="05cf5f68-7cfc-4caa-8db3-540b2c34caa0" containerID="73e8b74b9beec0b60b878c9d83c8a96f0567dc0ee44e26e1477d3ecd820cb721" exitCode=0 Oct 02 09:00:52 crc kubenswrapper[4829]: I1002 09:00:52.968470 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" event={"ID":"05cf5f68-7cfc-4caa-8db3-540b2c34caa0","Type":"ContainerDied","Data":"73e8b74b9beec0b60b878c9d83c8a96f0567dc0ee44e26e1477d3ecd820cb721"} Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.089577 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.213697 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnp2c\" (UniqueName: \"kubernetes.io/projected/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-kube-api-access-fnp2c\") pod \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.214048 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-host\") pod \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\" (UID: \"05cf5f68-7cfc-4caa-8db3-540b2c34caa0\") " Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.214139 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-host" (OuterVolumeSpecName: "host") pod "05cf5f68-7cfc-4caa-8db3-540b2c34caa0" (UID: "05cf5f68-7cfc-4caa-8db3-540b2c34caa0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.214861 4829 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.220147 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-kube-api-access-fnp2c" (OuterVolumeSpecName: "kube-api-access-fnp2c") pod "05cf5f68-7cfc-4caa-8db3-540b2c34caa0" (UID: "05cf5f68-7cfc-4caa-8db3-540b2c34caa0"). InnerVolumeSpecName "kube-api-access-fnp2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.316365 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnp2c\" (UniqueName: \"kubernetes.io/projected/05cf5f68-7cfc-4caa-8db3-540b2c34caa0-kube-api-access-fnp2c\") on node \"crc\" DevicePath \"\"" Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.987748 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" event={"ID":"05cf5f68-7cfc-4caa-8db3-540b2c34caa0","Type":"ContainerDied","Data":"d0b86e61a087599dc5591acbadc8852ed55cf49f2930f4ce19dad8bdaeb24508"} Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.987996 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0b86e61a087599dc5591acbadc8852ed55cf49f2930f4ce19dad8bdaeb24508" Oct 02 09:00:54 crc kubenswrapper[4829]: I1002 09:00:54.987781 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-hmjb7" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.140736 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29323261-4ws5r"] Oct 02 09:01:00 crc kubenswrapper[4829]: E1002 09:01:00.141752 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05cf5f68-7cfc-4caa-8db3-540b2c34caa0" containerName="container-00" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.141766 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="05cf5f68-7cfc-4caa-8db3-540b2c34caa0" containerName="container-00" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.141947 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="05cf5f68-7cfc-4caa-8db3-540b2c34caa0" containerName="container-00" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.142835 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.153440 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323261-4ws5r"] Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.213739 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-combined-ca-bundle\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.214090 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-config-data\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.214171 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tntt\" (UniqueName: \"kubernetes.io/projected/52679e92-937f-4995-bdc0-3ab20b068ac0-kube-api-access-8tntt\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.214270 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-fernet-keys\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.315725 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tntt\" (UniqueName: \"kubernetes.io/projected/52679e92-937f-4995-bdc0-3ab20b068ac0-kube-api-access-8tntt\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.315769 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-config-data\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.315797 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-fernet-keys\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.315826 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-combined-ca-bundle\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.321927 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-fernet-keys\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.322396 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-config-data\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.328028 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-combined-ca-bundle\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.337878 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tntt\" (UniqueName: \"kubernetes.io/projected/52679e92-937f-4995-bdc0-3ab20b068ac0-kube-api-access-8tntt\") pod \"keystone-cron-29323261-4ws5r\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.468481 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:00 crc kubenswrapper[4829]: I1002 09:01:00.937067 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29323261-4ws5r"] Oct 02 09:01:01 crc kubenswrapper[4829]: I1002 09:01:01.044595 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-4ws5r" event={"ID":"52679e92-937f-4995-bdc0-3ab20b068ac0","Type":"ContainerStarted","Data":"5e8ea797800dc4067827deb5bddee387ebe2324897897ce624c5122e8ed4609d"} Oct 02 09:01:01 crc kubenswrapper[4829]: I1002 09:01:01.993675 4829 scope.go:117] "RemoveContainer" containerID="ff7ebdfe8cdc5a0d46679c516037aab1aeb65178bf8de4a0976ab5e1c8e6b692" Oct 02 09:01:02 crc kubenswrapper[4829]: I1002 09:01:02.057181 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-4ws5r" event={"ID":"52679e92-937f-4995-bdc0-3ab20b068ac0","Type":"ContainerStarted","Data":"ed9222a990e20bd55e44c294546c63443719bfc7146e79f5e1995abe2ef4cfb5"} Oct 02 09:01:02 crc kubenswrapper[4829]: I1002 09:01:02.082914 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29323261-4ws5r" podStartSLOduration=2.082897107 podStartE2EDuration="2.082897107s" podCreationTimestamp="2025-10-02 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 09:01:02.073593637 +0000 UTC m=+6253.413242042" watchObservedRunningTime="2025-10-02 09:01:02.082897107 +0000 UTC m=+6253.422545512" Oct 02 09:01:02 crc kubenswrapper[4829]: I1002 09:01:02.118136 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-hmjb7"] Oct 02 09:01:02 crc kubenswrapper[4829]: I1002 09:01:02.126088 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-hmjb7"] Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.319342 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-k27vx"] Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.321349 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.323751 4829 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-qnkmf"/"default-dockercfg-2nbkl" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.381638 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d09bd758-9413-4eba-b486-c11d2460280c-host\") pod \"crc-debug-k27vx\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.381756 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt99j\" (UniqueName: \"kubernetes.io/projected/d09bd758-9413-4eba-b486-c11d2460280c-kube-api-access-gt99j\") pod \"crc-debug-k27vx\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.483444 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d09bd758-9413-4eba-b486-c11d2460280c-host\") pod \"crc-debug-k27vx\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.483554 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt99j\" (UniqueName: \"kubernetes.io/projected/d09bd758-9413-4eba-b486-c11d2460280c-kube-api-access-gt99j\") pod \"crc-debug-k27vx\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.483693 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d09bd758-9413-4eba-b486-c11d2460280c-host\") pod \"crc-debug-k27vx\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.484158 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05cf5f68-7cfc-4caa-8db3-540b2c34caa0" path="/var/lib/kubelet/pods/05cf5f68-7cfc-4caa-8db3-540b2c34caa0/volumes" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.512728 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt99j\" (UniqueName: \"kubernetes.io/projected/d09bd758-9413-4eba-b486-c11d2460280c-kube-api-access-gt99j\") pod \"crc-debug-k27vx\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: I1002 09:01:03.644126 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:03 crc kubenswrapper[4829]: W1002 09:01:03.688121 4829 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd09bd758_9413_4eba_b486_c11d2460280c.slice/crio-e2180e30e0dd115a67db7b2a37a2cadc2fb61988ff78bc514197807d77b49383 WatchSource:0}: Error finding container e2180e30e0dd115a67db7b2a37a2cadc2fb61988ff78bc514197807d77b49383: Status 404 returned error can't find the container with id e2180e30e0dd115a67db7b2a37a2cadc2fb61988ff78bc514197807d77b49383 Oct 02 09:01:04 crc kubenswrapper[4829]: I1002 09:01:04.082641 4829 generic.go:334] "Generic (PLEG): container finished" podID="52679e92-937f-4995-bdc0-3ab20b068ac0" containerID="ed9222a990e20bd55e44c294546c63443719bfc7146e79f5e1995abe2ef4cfb5" exitCode=0 Oct 02 09:01:04 crc kubenswrapper[4829]: I1002 09:01:04.082738 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-4ws5r" event={"ID":"52679e92-937f-4995-bdc0-3ab20b068ac0","Type":"ContainerDied","Data":"ed9222a990e20bd55e44c294546c63443719bfc7146e79f5e1995abe2ef4cfb5"} Oct 02 09:01:04 crc kubenswrapper[4829]: I1002 09:01:04.085423 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-k27vx" event={"ID":"d09bd758-9413-4eba-b486-c11d2460280c","Type":"ContainerStarted","Data":"af1995704d5d3486048373712e1aab49162d418001c7a9743f028a96467f4d20"} Oct 02 09:01:04 crc kubenswrapper[4829]: I1002 09:01:04.085466 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/crc-debug-k27vx" event={"ID":"d09bd758-9413-4eba-b486-c11d2460280c","Type":"ContainerStarted","Data":"e2180e30e0dd115a67db7b2a37a2cadc2fb61988ff78bc514197807d77b49383"} Oct 02 09:01:04 crc kubenswrapper[4829]: I1002 09:01:04.140193 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-k27vx"] Oct 02 09:01:04 crc kubenswrapper[4829]: I1002 09:01:04.151836 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qnkmf/crc-debug-k27vx"] Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.096256 4829 generic.go:334] "Generic (PLEG): container finished" podID="d09bd758-9413-4eba-b486-c11d2460280c" containerID="af1995704d5d3486048373712e1aab49162d418001c7a9743f028a96467f4d20" exitCode=0 Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.219235 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.318616 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d09bd758-9413-4eba-b486-c11d2460280c-host\") pod \"d09bd758-9413-4eba-b486-c11d2460280c\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.318744 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt99j\" (UniqueName: \"kubernetes.io/projected/d09bd758-9413-4eba-b486-c11d2460280c-kube-api-access-gt99j\") pod \"d09bd758-9413-4eba-b486-c11d2460280c\" (UID: \"d09bd758-9413-4eba-b486-c11d2460280c\") " Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.319397 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d09bd758-9413-4eba-b486-c11d2460280c-host" (OuterVolumeSpecName: "host") pod "d09bd758-9413-4eba-b486-c11d2460280c" (UID: "d09bd758-9413-4eba-b486-c11d2460280c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.326164 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d09bd758-9413-4eba-b486-c11d2460280c-kube-api-access-gt99j" (OuterVolumeSpecName: "kube-api-access-gt99j") pod "d09bd758-9413-4eba-b486-c11d2460280c" (UID: "d09bd758-9413-4eba-b486-c11d2460280c"). InnerVolumeSpecName "kube-api-access-gt99j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.421773 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt99j\" (UniqueName: \"kubernetes.io/projected/d09bd758-9413-4eba-b486-c11d2460280c-kube-api-access-gt99j\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.421815 4829 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d09bd758-9413-4eba-b486-c11d2460280c-host\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.472175 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d09bd758-9413-4eba-b486-c11d2460280c" path="/var/lib/kubelet/pods/d09bd758-9413-4eba-b486-c11d2460280c/volumes" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.504270 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.627663 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-combined-ca-bundle\") pod \"52679e92-937f-4995-bdc0-3ab20b068ac0\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.628034 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tntt\" (UniqueName: \"kubernetes.io/projected/52679e92-937f-4995-bdc0-3ab20b068ac0-kube-api-access-8tntt\") pod \"52679e92-937f-4995-bdc0-3ab20b068ac0\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.628128 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-config-data\") pod \"52679e92-937f-4995-bdc0-3ab20b068ac0\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.628200 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-fernet-keys\") pod \"52679e92-937f-4995-bdc0-3ab20b068ac0\" (UID: \"52679e92-937f-4995-bdc0-3ab20b068ac0\") " Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.635124 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "52679e92-937f-4995-bdc0-3ab20b068ac0" (UID: "52679e92-937f-4995-bdc0-3ab20b068ac0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.635952 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52679e92-937f-4995-bdc0-3ab20b068ac0-kube-api-access-8tntt" (OuterVolumeSpecName: "kube-api-access-8tntt") pod "52679e92-937f-4995-bdc0-3ab20b068ac0" (UID: "52679e92-937f-4995-bdc0-3ab20b068ac0"). InnerVolumeSpecName "kube-api-access-8tntt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.673978 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52679e92-937f-4995-bdc0-3ab20b068ac0" (UID: "52679e92-937f-4995-bdc0-3ab20b068ac0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.693840 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-config-data" (OuterVolumeSpecName: "config-data") pod "52679e92-937f-4995-bdc0-3ab20b068ac0" (UID: "52679e92-937f-4995-bdc0-3ab20b068ac0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.730590 4829 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.730628 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tntt\" (UniqueName: \"kubernetes.io/projected/52679e92-937f-4995-bdc0-3ab20b068ac0-kube-api-access-8tntt\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.730643 4829 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.730652 4829 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/52679e92-937f-4995-bdc0-3ab20b068ac0-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 09:01:05 crc kubenswrapper[4829]: I1002 09:01:05.863785 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/util/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.035152 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/pull/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.061396 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/util/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.061739 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/pull/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.111169 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29323261-4ws5r" event={"ID":"52679e92-937f-4995-bdc0-3ab20b068ac0","Type":"ContainerDied","Data":"5e8ea797800dc4067827deb5bddee387ebe2324897897ce624c5122e8ed4609d"} Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.111213 4829 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e8ea797800dc4067827deb5bddee387ebe2324897897ce624c5122e8ed4609d" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.111289 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29323261-4ws5r" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.116957 4829 scope.go:117] "RemoveContainer" containerID="af1995704d5d3486048373712e1aab49162d418001c7a9743f028a96467f4d20" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.117024 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/crc-debug-k27vx" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.484847 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/pull/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.500130 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/util/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.503043 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_92a9b6584006b4732453dd34e461a08cb03335466e56e0bd9de2f376b7jz4z5_1c06d3e9-2ef1-4e91-b24d-89cfeecf485e/extract/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.671050 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-nf7gj_895426fb-40c5-426b-ab48-f5ac31d22e06/kube-rbac-proxy/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.745621 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-nf7gj_895426fb-40c5-426b-ab48-f5ac31d22e06/manager/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.751903 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-7wrzt_f588d058-d57b-4824-b59d-40731650907f/kube-rbac-proxy/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.878890 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-7wrzt_f588d058-d57b-4824-b59d-40731650907f/manager/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.957340 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-rs8gs_349c96b5-e59f-4eac-8646-e5459775391c/manager/0.log" Oct 02 09:01:06 crc kubenswrapper[4829]: I1002 09:01:06.962688 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-rs8gs_349c96b5-e59f-4eac-8646-e5459775391c/kube-rbac-proxy/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.131469 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-5r5zb_104543c5-16f7-47d1-b2cb-dedd729a87b7/kube-rbac-proxy/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.214259 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-5r5zb_104543c5-16f7-47d1-b2cb-dedd729a87b7/manager/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.287950 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-ztngc_6f685eab-a458-4058-9453-5c4a389758e0/kube-rbac-proxy/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.315385 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-ztngc_6f685eab-a458-4058-9453-5c4a389758e0/manager/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.375295 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-cw4k8_5607af74-7a80-4b1c-9b33-690d3f7fe974/kube-rbac-proxy/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.471930 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-cw4k8_5607af74-7a80-4b1c-9b33-690d3f7fe974/manager/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.545895 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-zc9kp_e46a78e2-3b2a-47c5-8a9a-03877339340d/kube-rbac-proxy/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.730180 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-zc9kp_e46a78e2-3b2a-47c5-8a9a-03877339340d/manager/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.748603 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-2njwr_3f4b21a4-b546-4f69-a432-05ba29b24085/manager/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.751203 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-2njwr_3f4b21a4-b546-4f69-a432-05ba29b24085/kube-rbac-proxy/0.log" Oct 02 09:01:07 crc kubenswrapper[4829]: I1002 09:01:07.922722 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pf5r7_88cfec3f-d6c0-4007-a845-a3c4e56bcb63/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.015284 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-pf5r7_88cfec3f-d6c0-4007-a845-a3c4e56bcb63/manager/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.136489 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-tmszn_c555f81b-a5f9-4b4b-9ae0-5253ebb12099/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.150476 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-tmszn_c555f81b-a5f9-4b4b-9ae0-5253ebb12099/manager/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.218346 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-p8wc8_3430c81b-20da-4871-9f8c-a408573677ec/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.339174 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-p8wc8_3430c81b-20da-4871-9f8c-a408573677ec/manager/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.381656 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-2rk2r_d697f00d-25f4-48fb-808e-1a5fb90ff5fa/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.457515 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-2rk2r_d697f00d-25f4-48fb-808e-1a5fb90ff5fa/manager/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.526100 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-722dn_bf88c28f-c14a-4390-83ed-75dc3e41061d/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.643258 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-722dn_bf88c28f-c14a-4390-83ed-75dc3e41061d/manager/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.763360 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-hlq5c_c35efaf9-0b1b-4801-81c9-f8db24083049/manager/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.786641 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-hlq5c_c35efaf9-0b1b-4801-81c9-f8db24083049/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.873951 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-swsvz_2ba71fca-3e78-49e5-bfe5-b831568118ba/kube-rbac-proxy/0.log" Oct 02 09:01:08 crc kubenswrapper[4829]: I1002 09:01:08.958053 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-swsvz_2ba71fca-3e78-49e5-bfe5-b831568118ba/manager/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.068074 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f874976-r72h2_f3a3eba9-d7d0-4796-acdc-2b9525a17e7a/kube-rbac-proxy/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.215328 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6f46d87668-x4x8h_507314d1-5e67-4900-99bd-fd445866f35b/kube-rbac-proxy/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.448481 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6f46d87668-x4x8h_507314d1-5e67-4900-99bd-fd445866f35b/operator/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.530371 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-bqsd6_fba8d39c-ec14-48e6-a9de-41f33e873f09/registry-server/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.781094 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-5cqsw_3dee869a-23fe-4926-a561-67a87ca83103/kube-rbac-proxy/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.790254 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-5cqsw_3dee869a-23fe-4926-a561-67a87ca83103/manager/0.log" Oct 02 09:01:09 crc kubenswrapper[4829]: I1002 09:01:09.888385 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-vv8cw_8dd30f73-72f3-4733-b5c8-f96753750fb8/kube-rbac-proxy/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.078570 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-vv8cw_8dd30f73-72f3-4733-b5c8-f96753750fb8/manager/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.170900 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-d4t94_1c017022-5036-4312-b914-8aeb5439a9ef/operator/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.230530 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79f874976-r72h2_f3a3eba9-d7d0-4796-acdc-2b9525a17e7a/manager/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.293508 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-mqflt_fb49bfb5-3c2c-45eb-94d5-8061c919435d/kube-rbac-proxy/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.334906 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-mqflt_fb49bfb5-3c2c-45eb-94d5-8061c919435d/manager/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.376054 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-4wdcm_71170543-3bd9-4d9f-9ad4-a5978a6f018b/kube-rbac-proxy/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.491783 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-twxn6_8129c7b9-335b-4bda-8516-94a818a57591/kube-rbac-proxy/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.603203 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-twxn6_8129c7b9-335b-4bda-8516-94a818a57591/manager/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.610140 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-4wdcm_71170543-3bd9-4d9f-9ad4-a5978a6f018b/manager/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.867560 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6bf8759486-hrl49_a3ae11b8-1d62-41ba-a63b-4441d2f70709/kube-rbac-proxy/0.log" Oct 02 09:01:10 crc kubenswrapper[4829]: I1002 09:01:10.974407 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6bf8759486-hrl49_a3ae11b8-1d62-41ba-a63b-4441d2f70709/manager/0.log" Oct 02 09:01:25 crc kubenswrapper[4829]: I1002 09:01:25.329849 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:01:25 crc kubenswrapper[4829]: I1002 09:01:25.330388 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:01:28 crc kubenswrapper[4829]: I1002 09:01:28.197902 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zqd5d_2cc39185-906e-4b18-a359-b1fd011b74f4/control-plane-machine-set-operator/0.log" Oct 02 09:01:28 crc kubenswrapper[4829]: I1002 09:01:28.362440 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hwmtz_6057de60-fc56-49b5-8843-a9e838186747/machine-api-operator/0.log" Oct 02 09:01:28 crc kubenswrapper[4829]: I1002 09:01:28.362697 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-hwmtz_6057de60-fc56-49b5-8843-a9e838186747/kube-rbac-proxy/0.log" Oct 02 09:01:40 crc kubenswrapper[4829]: I1002 09:01:40.905788 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-d6q7t_482724e1-76a1-40b1-b5e2-1a51cade2778/cert-manager-controller/0.log" Oct 02 09:01:40 crc kubenswrapper[4829]: I1002 09:01:40.978554 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-tc7zk_13d4846f-c8e7-40a4-b312-3b3236539721/cert-manager-cainjector/0.log" Oct 02 09:01:41 crc kubenswrapper[4829]: I1002 09:01:41.025334 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-f6xpf_adcd7359-41d5-433b-a4d1-aa964b09f997/cert-manager-webhook/0.log" Oct 02 09:01:53 crc kubenswrapper[4829]: I1002 09:01:53.593994 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-t5rvr_d9eea7b1-65ab-4699-ad35-15aad5a523f9/nmstate-console-plugin/0.log" Oct 02 09:01:53 crc kubenswrapper[4829]: I1002 09:01:53.738208 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kgg6f_6b7faad3-6c8e-4cee-a257-f8bc8ef83e0e/nmstate-handler/0.log" Oct 02 09:01:53 crc kubenswrapper[4829]: I1002 09:01:53.787719 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8rhft_3795be47-c66b-4804-b130-33a73c5e57ce/kube-rbac-proxy/0.log" Oct 02 09:01:53 crc kubenswrapper[4829]: I1002 09:01:53.801186 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8rhft_3795be47-c66b-4804-b130-33a73c5e57ce/nmstate-metrics/0.log" Oct 02 09:01:53 crc kubenswrapper[4829]: I1002 09:01:53.949237 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-hl4ld_7bc8f7d2-12dc-43a9-8cae-872d39a4defc/nmstate-operator/0.log" Oct 02 09:01:54 crc kubenswrapper[4829]: I1002 09:01:54.014445 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-6q26v_7e10a126-c321-4d0b-9757-080946eddbba/nmstate-webhook/0.log" Oct 02 09:01:55 crc kubenswrapper[4829]: I1002 09:01:55.329076 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:01:55 crc kubenswrapper[4829]: I1002 09:01:55.329510 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:02:07 crc kubenswrapper[4829]: I1002 09:02:07.957514 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-mpn22_321a9e53-9458-4d2f-9784-15f96c7d157b/kube-rbac-proxy/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.124759 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.148114 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-mpn22_321a9e53-9458-4d2f-9784-15f96c7d157b/controller/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.333578 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.383056 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.388203 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.426414 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.602124 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.609364 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.611697 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.632558 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.783582 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/controller/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.790586 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-reloader/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.800940 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-frr-files/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.819373 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/cp-metrics/0.log" Oct 02 09:02:08 crc kubenswrapper[4829]: I1002 09:02:08.980523 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/kube-rbac-proxy-frr/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.016370 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/kube-rbac-proxy/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.018783 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/frr-metrics/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.180044 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/reloader/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.222815 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-t78d2_99893ca5-1605-4a55-9ce8-6711e86fa31b/frr-k8s-webhook-server/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.490497 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-75cb48855d-jh2hq_65093a97-cc2b-4983-a106-60b3af7783ca/manager/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.619323 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6884746998-tvx77_67cefc41-a1b6-4c77-9770-ea2c4f0fae83/webhook-server/0.log" Oct 02 09:02:09 crc kubenswrapper[4829]: I1002 09:02:09.733027 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mvlxd_3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64/kube-rbac-proxy/0.log" Oct 02 09:02:10 crc kubenswrapper[4829]: I1002 09:02:10.340459 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-mvlxd_3af56f1e-7c39-40ef-bf6a-ee5ccaaeda64/speaker/0.log" Oct 02 09:02:10 crc kubenswrapper[4829]: I1002 09:02:10.688149 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8sxd9_a85bf9a0-7c63-492e-8662-c463be70aad0/frr/0.log" Oct 02 09:02:22 crc kubenswrapper[4829]: I1002 09:02:22.960505 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/util/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.176284 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/util/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.235451 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/pull/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.250413 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/pull/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.382151 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/util/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.421953 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/pull/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.422207 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2drlg7_036e78a8-b2cb-4210-bbfd-dedba601b30a/extract/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.601476 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/util/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.759190 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/util/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.773019 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/pull/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.787451 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/pull/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.922316 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/util/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.929370 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/extract/0.log" Oct 02 09:02:23 crc kubenswrapper[4829]: I1002 09:02:23.946607 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dl746w_ff58a0e1-dfff-438b-bd40-00240f7372dd/pull/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.097416 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-utilities/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.249862 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-content/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.252289 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-utilities/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.290018 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-content/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.429084 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-content/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.429409 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/extract-utilities/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.650954 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-utilities/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.881264 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-utilities/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.891387 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-content/0.log" Oct 02 09:02:24 crc kubenswrapper[4829]: I1002 09:02:24.948829 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-content/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.109213 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-pv5qf_866eefa0-245f-4ed1-b93c-000cf03733b0/registry-server/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.154665 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-utilities/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.155359 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/extract-content/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.328644 4829 patch_prober.go:28] interesting pod/machine-config-daemon-8rvq4 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.328693 4829 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.328738 4829 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.329445 4829 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c"} pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.329496 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerName="machine-config-daemon" containerID="cri-o://638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" gracePeriod=600 Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.374900 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/util/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: E1002 09:02:25.452474 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.695667 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/pull/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.709719 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/pull/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.716396 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/util/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.894989 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/pull/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.915962 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/extract/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.928363 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7t9nm_d840464b-b486-4421-acd8-fded53601478/util/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.965749 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-n95gp_d0590b92-f618-4e54-aea6-66c88f9f0f4a/registry-server/0.log" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.982932 4829 generic.go:334] "Generic (PLEG): container finished" podID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" exitCode=0 Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.982970 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" event={"ID":"60aa0a81-354e-46f1-ab0c-b1eb386974a6","Type":"ContainerDied","Data":"638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c"} Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.983007 4829 scope.go:117] "RemoveContainer" containerID="64097f1b48a317b51134dfe4c97d1b7feba293cedd22314bc31de6b95e27b269" Oct 02 09:02:25 crc kubenswrapper[4829]: I1002 09:02:25.983652 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:02:25 crc kubenswrapper[4829]: E1002 09:02:25.983908 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.135309 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-bbdf8_156b7e29-6cb2-4979-9ef3-c9f7904ecc0f/marketplace-operator/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.166005 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-utilities/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.326039 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-utilities/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.331345 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-content/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.362830 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-content/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.484845 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-utilities/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.552663 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/extract-content/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.603817 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-utilities/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.704601 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-9spqb_cef47a6d-4a19-45e0-9dca-2ab3a4db27dc/registry-server/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.785200 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-content/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.787246 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-utilities/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.857637 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-content/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.970521 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-utilities/0.log" Oct 02 09:02:26 crc kubenswrapper[4829]: I1002 09:02:26.988590 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/extract-content/0.log" Oct 02 09:02:27 crc kubenswrapper[4829]: I1002 09:02:27.660886 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-pc6x5_b64a03ff-edef-475b-ae70-739d43cefd48/registry-server/0.log" Oct 02 09:02:39 crc kubenswrapper[4829]: I1002 09:02:39.666220 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-448hn_873d1771-f008-4ac1-8ece-e3b48ba37a81/prometheus-operator/0.log" Oct 02 09:02:39 crc kubenswrapper[4829]: I1002 09:02:39.850233 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7bf9c989b9-g9kcz_c1ddcd62-beba-4339-9aa1-df07ce7577a8/prometheus-operator-admission-webhook/0.log" Oct 02 09:02:39 crc kubenswrapper[4829]: I1002 09:02:39.856858 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7bf9c989b9-cs9mv_9f5f2c2d-e40c-4b14-97df-a5a83489726a/prometheus-operator-admission-webhook/0.log" Oct 02 09:02:40 crc kubenswrapper[4829]: I1002 09:02:40.013681 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-6th2l_9175ba03-d3bd-4278-8943-1499754f510b/operator/0.log" Oct 02 09:02:40 crc kubenswrapper[4829]: I1002 09:02:40.056949 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-mw7fx_eb1be399-f44c-4acb-81ad-9723071d7f9f/perses-operator/0.log" Oct 02 09:02:40 crc kubenswrapper[4829]: I1002 09:02:40.461096 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:02:40 crc kubenswrapper[4829]: E1002 09:02:40.461397 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.262107 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m5mrm"] Oct 02 09:02:41 crc kubenswrapper[4829]: E1002 09:02:41.263042 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52679e92-937f-4995-bdc0-3ab20b068ac0" containerName="keystone-cron" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.263063 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="52679e92-937f-4995-bdc0-3ab20b068ac0" containerName="keystone-cron" Oct 02 09:02:41 crc kubenswrapper[4829]: E1002 09:02:41.263110 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d09bd758-9413-4eba-b486-c11d2460280c" containerName="container-00" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.263122 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="d09bd758-9413-4eba-b486-c11d2460280c" containerName="container-00" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.263642 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="52679e92-937f-4995-bdc0-3ab20b068ac0" containerName="keystone-cron" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.263685 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="d09bd758-9413-4eba-b486-c11d2460280c" containerName="container-00" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.268436 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.291024 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5mrm"] Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.363154 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-466zw\" (UniqueName: \"kubernetes.io/projected/aec41807-47d3-4595-b63b-c879dd400e58-kube-api-access-466zw\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.363283 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-utilities\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.363351 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-catalog-content\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.464696 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-catalog-content\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.464832 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-466zw\" (UniqueName: \"kubernetes.io/projected/aec41807-47d3-4595-b63b-c879dd400e58-kube-api-access-466zw\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.464909 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-utilities\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.465308 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-catalog-content\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.465401 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-utilities\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.508005 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-466zw\" (UniqueName: \"kubernetes.io/projected/aec41807-47d3-4595-b63b-c879dd400e58-kube-api-access-466zw\") pod \"certified-operators-m5mrm\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:41 crc kubenswrapper[4829]: I1002 09:02:41.606534 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:42 crc kubenswrapper[4829]: I1002 09:02:42.102595 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5mrm"] Oct 02 09:02:42 crc kubenswrapper[4829]: I1002 09:02:42.153694 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerStarted","Data":"5876b45bca966630586e0796b4a9c84172e395140ce314c8e057df5a8d5b1a56"} Oct 02 09:02:43 crc kubenswrapper[4829]: I1002 09:02:43.164180 4829 generic.go:334] "Generic (PLEG): container finished" podID="aec41807-47d3-4595-b63b-c879dd400e58" containerID="9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04" exitCode=0 Oct 02 09:02:43 crc kubenswrapper[4829]: I1002 09:02:43.164216 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerDied","Data":"9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04"} Oct 02 09:02:44 crc kubenswrapper[4829]: I1002 09:02:44.175245 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerStarted","Data":"ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950"} Oct 02 09:02:46 crc kubenswrapper[4829]: I1002 09:02:46.200475 4829 generic.go:334] "Generic (PLEG): container finished" podID="aec41807-47d3-4595-b63b-c879dd400e58" containerID="ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950" exitCode=0 Oct 02 09:02:46 crc kubenswrapper[4829]: I1002 09:02:46.200801 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerDied","Data":"ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950"} Oct 02 09:02:47 crc kubenswrapper[4829]: I1002 09:02:47.214630 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerStarted","Data":"8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03"} Oct 02 09:02:47 crc kubenswrapper[4829]: I1002 09:02:47.239726 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m5mrm" podStartSLOduration=2.695519783 podStartE2EDuration="6.239704144s" podCreationTimestamp="2025-10-02 09:02:41 +0000 UTC" firstStartedPulling="2025-10-02 09:02:43.166444872 +0000 UTC m=+6354.506093287" lastFinishedPulling="2025-10-02 09:02:46.710629243 +0000 UTC m=+6358.050277648" observedRunningTime="2025-10-02 09:02:47.232266542 +0000 UTC m=+6358.571914957" watchObservedRunningTime="2025-10-02 09:02:47.239704144 +0000 UTC m=+6358.579352559" Oct 02 09:02:51 crc kubenswrapper[4829]: I1002 09:02:51.606894 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:51 crc kubenswrapper[4829]: I1002 09:02:51.607296 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:02:52 crc kubenswrapper[4829]: I1002 09:02:52.677273 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-m5mrm" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="registry-server" probeResult="failure" output=< Oct 02 09:02:52 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 09:02:52 crc kubenswrapper[4829]: > Oct 02 09:02:54 crc kubenswrapper[4829]: I1002 09:02:54.460456 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:02:54 crc kubenswrapper[4829]: E1002 09:02:54.461402 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:03:01 crc kubenswrapper[4829]: I1002 09:03:01.660622 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:03:01 crc kubenswrapper[4829]: I1002 09:03:01.734441 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:03:01 crc kubenswrapper[4829]: I1002 09:03:01.900721 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5mrm"] Oct 02 09:03:03 crc kubenswrapper[4829]: I1002 09:03:03.380000 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m5mrm" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="registry-server" containerID="cri-o://8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03" gracePeriod=2 Oct 02 09:03:03 crc kubenswrapper[4829]: I1002 09:03:03.938220 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.041704 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-utilities\") pod \"aec41807-47d3-4595-b63b-c879dd400e58\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.041763 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-466zw\" (UniqueName: \"kubernetes.io/projected/aec41807-47d3-4595-b63b-c879dd400e58-kube-api-access-466zw\") pod \"aec41807-47d3-4595-b63b-c879dd400e58\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.041982 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-catalog-content\") pod \"aec41807-47d3-4595-b63b-c879dd400e58\" (UID: \"aec41807-47d3-4595-b63b-c879dd400e58\") " Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.043152 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-utilities" (OuterVolumeSpecName: "utilities") pod "aec41807-47d3-4595-b63b-c879dd400e58" (UID: "aec41807-47d3-4595-b63b-c879dd400e58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.049131 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec41807-47d3-4595-b63b-c879dd400e58-kube-api-access-466zw" (OuterVolumeSpecName: "kube-api-access-466zw") pod "aec41807-47d3-4595-b63b-c879dd400e58" (UID: "aec41807-47d3-4595-b63b-c879dd400e58"). InnerVolumeSpecName "kube-api-access-466zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.056403 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.056452 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-466zw\" (UniqueName: \"kubernetes.io/projected/aec41807-47d3-4595-b63b-c879dd400e58-kube-api-access-466zw\") on node \"crc\" DevicePath \"\"" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.086246 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aec41807-47d3-4595-b63b-c879dd400e58" (UID: "aec41807-47d3-4595-b63b-c879dd400e58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.157836 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec41807-47d3-4595-b63b-c879dd400e58-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.395374 4829 generic.go:334] "Generic (PLEG): container finished" podID="aec41807-47d3-4595-b63b-c879dd400e58" containerID="8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03" exitCode=0 Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.395437 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerDied","Data":"8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03"} Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.395478 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5mrm" event={"ID":"aec41807-47d3-4595-b63b-c879dd400e58","Type":"ContainerDied","Data":"5876b45bca966630586e0796b4a9c84172e395140ce314c8e057df5a8d5b1a56"} Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.395504 4829 scope.go:117] "RemoveContainer" containerID="8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.395592 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5mrm" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.427170 4829 scope.go:117] "RemoveContainer" containerID="ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.504949 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5mrm"] Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.531436 4829 scope.go:117] "RemoveContainer" containerID="9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.532195 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m5mrm"] Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.563382 4829 scope.go:117] "RemoveContainer" containerID="8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03" Oct 02 09:03:04 crc kubenswrapper[4829]: E1002 09:03:04.564332 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03\": container with ID starting with 8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03 not found: ID does not exist" containerID="8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.564393 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03"} err="failed to get container status \"8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03\": rpc error: code = NotFound desc = could not find container \"8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03\": container with ID starting with 8f5850b58f2aa1d3c7e79a0bf28d29e9494f542474d3a0bfb2c08e1a969f0c03 not found: ID does not exist" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.564433 4829 scope.go:117] "RemoveContainer" containerID="ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950" Oct 02 09:03:04 crc kubenswrapper[4829]: E1002 09:03:04.564970 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950\": container with ID starting with ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950 not found: ID does not exist" containerID="ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.565175 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950"} err="failed to get container status \"ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950\": rpc error: code = NotFound desc = could not find container \"ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950\": container with ID starting with ff8e7e55f03fc54bc583bf223c6ac81cf8392f43a9b78056e847fc10d9e63950 not found: ID does not exist" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.565390 4829 scope.go:117] "RemoveContainer" containerID="9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04" Oct 02 09:03:04 crc kubenswrapper[4829]: E1002 09:03:04.566136 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04\": container with ID starting with 9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04 not found: ID does not exist" containerID="9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04" Oct 02 09:03:04 crc kubenswrapper[4829]: I1002 09:03:04.566191 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04"} err="failed to get container status \"9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04\": rpc error: code = NotFound desc = could not find container \"9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04\": container with ID starting with 9388ebf23a7e6ad5f917cb2f79c9f6f8cd69f4167cabe3b534fd36ea6ab27a04 not found: ID does not exist" Oct 02 09:03:05 crc kubenswrapper[4829]: I1002 09:03:05.483897 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aec41807-47d3-4595-b63b-c879dd400e58" path="/var/lib/kubelet/pods/aec41807-47d3-4595-b63b-c879dd400e58/volumes" Oct 02 09:03:07 crc kubenswrapper[4829]: I1002 09:03:07.462103 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:03:07 crc kubenswrapper[4829]: E1002 09:03:07.462961 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:03:20 crc kubenswrapper[4829]: I1002 09:03:20.461967 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:03:20 crc kubenswrapper[4829]: E1002 09:03:20.463424 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:03:34 crc kubenswrapper[4829]: I1002 09:03:34.461201 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:03:34 crc kubenswrapper[4829]: E1002 09:03:34.463642 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:03:48 crc kubenswrapper[4829]: I1002 09:03:48.461966 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:03:48 crc kubenswrapper[4829]: E1002 09:03:48.463061 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:04:01 crc kubenswrapper[4829]: I1002 09:04:01.461407 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:04:01 crc kubenswrapper[4829]: E1002 09:04:01.462076 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:04:16 crc kubenswrapper[4829]: I1002 09:04:16.462810 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:04:16 crc kubenswrapper[4829]: E1002 09:04:16.463705 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:04:27 crc kubenswrapper[4829]: I1002 09:04:27.463340 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:04:27 crc kubenswrapper[4829]: E1002 09:04:27.464372 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:04:40 crc kubenswrapper[4829]: I1002 09:04:40.462274 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:04:40 crc kubenswrapper[4829]: E1002 09:04:40.463429 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:04:49 crc kubenswrapper[4829]: I1002 09:04:49.683529 4829 generic.go:334] "Generic (PLEG): container finished" podID="10a628e6-642f-4f9c-9234-809aecaa94de" containerID="8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336" exitCode=0 Oct 02 09:04:49 crc kubenswrapper[4829]: I1002 09:04:49.683955 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-qnkmf/must-gather-9nssl" event={"ID":"10a628e6-642f-4f9c-9234-809aecaa94de","Type":"ContainerDied","Data":"8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336"} Oct 02 09:04:49 crc kubenswrapper[4829]: I1002 09:04:49.684666 4829 scope.go:117] "RemoveContainer" containerID="8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336" Oct 02 09:04:49 crc kubenswrapper[4829]: I1002 09:04:49.799816 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qnkmf_must-gather-9nssl_10a628e6-642f-4f9c-9234-809aecaa94de/gather/0.log" Oct 02 09:04:53 crc kubenswrapper[4829]: I1002 09:04:53.462036 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:04:53 crc kubenswrapper[4829]: E1002 09:04:53.462979 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:04:56 crc kubenswrapper[4829]: E1002 09:04:56.743711 4829 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.200:55490->38.102.83.200:43623: write tcp 38.102.83.200:55490->38.102.83.200:43623: write: broken pipe Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.672598 4829 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n95vq"] Oct 02 09:05:01 crc kubenswrapper[4829]: E1002 09:05:01.673784 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="registry-server" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.673801 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="registry-server" Oct 02 09:05:01 crc kubenswrapper[4829]: E1002 09:05:01.673849 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="extract-utilities" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.673869 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="extract-utilities" Oct 02 09:05:01 crc kubenswrapper[4829]: E1002 09:05:01.673894 4829 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="extract-content" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.673903 4829 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="extract-content" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.674164 4829 memory_manager.go:354] "RemoveStaleState removing state" podUID="aec41807-47d3-4595-b63b-c879dd400e58" containerName="registry-server" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.676049 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.703358 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n95vq"] Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.794561 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-catalog-content\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.794695 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwqk2\" (UniqueName: \"kubernetes.io/projected/6a98eac3-f47e-4b62-991d-9031531884cb-kube-api-access-fwqk2\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.794783 4829 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-utilities\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.897131 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwqk2\" (UniqueName: \"kubernetes.io/projected/6a98eac3-f47e-4b62-991d-9031531884cb-kube-api-access-fwqk2\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.897265 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-utilities\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.897376 4829 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-catalog-content\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.897934 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-catalog-content\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.897952 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-utilities\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:01 crc kubenswrapper[4829]: I1002 09:05:01.917820 4829 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwqk2\" (UniqueName: \"kubernetes.io/projected/6a98eac3-f47e-4b62-991d-9031531884cb-kube-api-access-fwqk2\") pod \"redhat-operators-n95vq\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.000078 4829 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.193099 4829 scope.go:117] "RemoveContainer" containerID="44cf2611d099cf1e9e878bb9837af983a1f98791d07fcb7447b24d7cb68296e2" Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.491089 4829 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n95vq"] Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.827566 4829 generic.go:334] "Generic (PLEG): container finished" podID="6a98eac3-f47e-4b62-991d-9031531884cb" containerID="c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc" exitCode=0 Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.827608 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerDied","Data":"c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc"} Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.827627 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerStarted","Data":"e992f998aea24e30a8b8965b36838bb6d2e09bee1df00831be9f7eab630b39bc"} Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.829422 4829 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.868757 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-qnkmf/must-gather-9nssl"] Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.869236 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-qnkmf/must-gather-9nssl" podUID="10a628e6-642f-4f9c-9234-809aecaa94de" containerName="copy" containerID="cri-o://2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1" gracePeriod=2 Oct 02 09:05:02 crc kubenswrapper[4829]: I1002 09:05:02.879638 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-qnkmf/must-gather-9nssl"] Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.350614 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qnkmf_must-gather-9nssl_10a628e6-642f-4f9c-9234-809aecaa94de/copy/0.log" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.351176 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.431022 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10a628e6-642f-4f9c-9234-809aecaa94de-must-gather-output\") pod \"10a628e6-642f-4f9c-9234-809aecaa94de\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.431072 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spcnh\" (UniqueName: \"kubernetes.io/projected/10a628e6-642f-4f9c-9234-809aecaa94de-kube-api-access-spcnh\") pod \"10a628e6-642f-4f9c-9234-809aecaa94de\" (UID: \"10a628e6-642f-4f9c-9234-809aecaa94de\") " Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.441589 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10a628e6-642f-4f9c-9234-809aecaa94de-kube-api-access-spcnh" (OuterVolumeSpecName: "kube-api-access-spcnh") pod "10a628e6-642f-4f9c-9234-809aecaa94de" (UID: "10a628e6-642f-4f9c-9234-809aecaa94de"). InnerVolumeSpecName "kube-api-access-spcnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.533261 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spcnh\" (UniqueName: \"kubernetes.io/projected/10a628e6-642f-4f9c-9234-809aecaa94de-kube-api-access-spcnh\") on node \"crc\" DevicePath \"\"" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.633933 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10a628e6-642f-4f9c-9234-809aecaa94de-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "10a628e6-642f-4f9c-9234-809aecaa94de" (UID: "10a628e6-642f-4f9c-9234-809aecaa94de"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.634682 4829 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/10a628e6-642f-4f9c-9234-809aecaa94de-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.838776 4829 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-qnkmf_must-gather-9nssl_10a628e6-642f-4f9c-9234-809aecaa94de/copy/0.log" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.839295 4829 generic.go:334] "Generic (PLEG): container finished" podID="10a628e6-642f-4f9c-9234-809aecaa94de" containerID="2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1" exitCode=143 Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.839342 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-qnkmf/must-gather-9nssl" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.839371 4829 scope.go:117] "RemoveContainer" containerID="2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.868723 4829 scope.go:117] "RemoveContainer" containerID="8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.942266 4829 scope.go:117] "RemoveContainer" containerID="2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1" Oct 02 09:05:03 crc kubenswrapper[4829]: E1002 09:05:03.942645 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1\": container with ID starting with 2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1 not found: ID does not exist" containerID="2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.942691 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1"} err="failed to get container status \"2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1\": rpc error: code = NotFound desc = could not find container \"2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1\": container with ID starting with 2e7ab51e4573c02180c315acbe41947644bc4fae5f30170ec3c77ae40d11c8e1 not found: ID does not exist" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.942720 4829 scope.go:117] "RemoveContainer" containerID="8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336" Oct 02 09:05:03 crc kubenswrapper[4829]: E1002 09:05:03.942998 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336\": container with ID starting with 8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336 not found: ID does not exist" containerID="8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336" Oct 02 09:05:03 crc kubenswrapper[4829]: I1002 09:05:03.943024 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336"} err="failed to get container status \"8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336\": rpc error: code = NotFound desc = could not find container \"8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336\": container with ID starting with 8e259f36e7b7f368b76a5af51428e5b658266f060c221acc854e59ddeebb2336 not found: ID does not exist" Oct 02 09:05:04 crc kubenswrapper[4829]: I1002 09:05:04.460715 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:05:04 crc kubenswrapper[4829]: E1002 09:05:04.461330 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:05:04 crc kubenswrapper[4829]: I1002 09:05:04.850540 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerStarted","Data":"bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0"} Oct 02 09:05:05 crc kubenswrapper[4829]: I1002 09:05:05.478171 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10a628e6-642f-4f9c-9234-809aecaa94de" path="/var/lib/kubelet/pods/10a628e6-642f-4f9c-9234-809aecaa94de/volumes" Oct 02 09:05:07 crc kubenswrapper[4829]: I1002 09:05:07.885347 4829 generic.go:334] "Generic (PLEG): container finished" podID="6a98eac3-f47e-4b62-991d-9031531884cb" containerID="bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0" exitCode=0 Oct 02 09:05:07 crc kubenswrapper[4829]: I1002 09:05:07.885394 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerDied","Data":"bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0"} Oct 02 09:05:08 crc kubenswrapper[4829]: I1002 09:05:08.897833 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerStarted","Data":"989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2"} Oct 02 09:05:08 crc kubenswrapper[4829]: I1002 09:05:08.925991 4829 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n95vq" podStartSLOduration=2.444833815 podStartE2EDuration="7.925953659s" podCreationTimestamp="2025-10-02 09:05:01 +0000 UTC" firstStartedPulling="2025-10-02 09:05:02.829189666 +0000 UTC m=+6494.168838071" lastFinishedPulling="2025-10-02 09:05:08.31030946 +0000 UTC m=+6499.649957915" observedRunningTime="2025-10-02 09:05:08.912422707 +0000 UTC m=+6500.252071112" watchObservedRunningTime="2025-10-02 09:05:08.925953659 +0000 UTC m=+6500.265602084" Oct 02 09:05:12 crc kubenswrapper[4829]: I1002 09:05:12.001075 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:12 crc kubenswrapper[4829]: I1002 09:05:12.001682 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:13 crc kubenswrapper[4829]: I1002 09:05:13.048474 4829 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n95vq" podUID="6a98eac3-f47e-4b62-991d-9031531884cb" containerName="registry-server" probeResult="failure" output=< Oct 02 09:05:13 crc kubenswrapper[4829]: timeout: failed to connect service ":50051" within 1s Oct 02 09:05:13 crc kubenswrapper[4829]: > Oct 02 09:05:16 crc kubenswrapper[4829]: I1002 09:05:16.460881 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:05:16 crc kubenswrapper[4829]: E1002 09:05:16.461894 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:05:22 crc kubenswrapper[4829]: I1002 09:05:22.077674 4829 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:22 crc kubenswrapper[4829]: I1002 09:05:22.161286 4829 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:22 crc kubenswrapper[4829]: I1002 09:05:22.325198 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n95vq"] Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.048567 4829 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n95vq" podUID="6a98eac3-f47e-4b62-991d-9031531884cb" containerName="registry-server" containerID="cri-o://989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2" gracePeriod=2 Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.645165 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.768329 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwqk2\" (UniqueName: \"kubernetes.io/projected/6a98eac3-f47e-4b62-991d-9031531884cb-kube-api-access-fwqk2\") pod \"6a98eac3-f47e-4b62-991d-9031531884cb\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.768571 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-utilities\") pod \"6a98eac3-f47e-4b62-991d-9031531884cb\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.768610 4829 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-catalog-content\") pod \"6a98eac3-f47e-4b62-991d-9031531884cb\" (UID: \"6a98eac3-f47e-4b62-991d-9031531884cb\") " Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.769416 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-utilities" (OuterVolumeSpecName: "utilities") pod "6a98eac3-f47e-4b62-991d-9031531884cb" (UID: "6a98eac3-f47e-4b62-991d-9031531884cb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.773482 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a98eac3-f47e-4b62-991d-9031531884cb-kube-api-access-fwqk2" (OuterVolumeSpecName: "kube-api-access-fwqk2") pod "6a98eac3-f47e-4b62-991d-9031531884cb" (UID: "6a98eac3-f47e-4b62-991d-9031531884cb"). InnerVolumeSpecName "kube-api-access-fwqk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.871314 4829 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.871346 4829 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwqk2\" (UniqueName: \"kubernetes.io/projected/6a98eac3-f47e-4b62-991d-9031531884cb-kube-api-access-fwqk2\") on node \"crc\" DevicePath \"\"" Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.875404 4829 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a98eac3-f47e-4b62-991d-9031531884cb" (UID: "6a98eac3-f47e-4b62-991d-9031531884cb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 09:05:24 crc kubenswrapper[4829]: I1002 09:05:24.972625 4829 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a98eac3-f47e-4b62-991d-9031531884cb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.059869 4829 generic.go:334] "Generic (PLEG): container finished" podID="6a98eac3-f47e-4b62-991d-9031531884cb" containerID="989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2" exitCode=0 Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.059912 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerDied","Data":"989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2"} Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.059947 4829 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n95vq" event={"ID":"6a98eac3-f47e-4b62-991d-9031531884cb","Type":"ContainerDied","Data":"e992f998aea24e30a8b8965b36838bb6d2e09bee1df00831be9f7eab630b39bc"} Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.059978 4829 scope.go:117] "RemoveContainer" containerID="989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.060887 4829 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n95vq" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.080013 4829 scope.go:117] "RemoveContainer" containerID="bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.097545 4829 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n95vq"] Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.108109 4829 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n95vq"] Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.119366 4829 scope.go:117] "RemoveContainer" containerID="c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.172771 4829 scope.go:117] "RemoveContainer" containerID="989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2" Oct 02 09:05:25 crc kubenswrapper[4829]: E1002 09:05:25.173175 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2\": container with ID starting with 989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2 not found: ID does not exist" containerID="989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.173213 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2"} err="failed to get container status \"989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2\": rpc error: code = NotFound desc = could not find container \"989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2\": container with ID starting with 989c61f388bb71a3740f6765cbd93abdc69caf48cf09e3c3db3811d6628d71e2 not found: ID does not exist" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.173255 4829 scope.go:117] "RemoveContainer" containerID="bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0" Oct 02 09:05:25 crc kubenswrapper[4829]: E1002 09:05:25.173506 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0\": container with ID starting with bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0 not found: ID does not exist" containerID="bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.173535 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0"} err="failed to get container status \"bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0\": rpc error: code = NotFound desc = could not find container \"bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0\": container with ID starting with bc1f110be452231bae86198b031954b54aaa7c2c3728fb7560fdb67a2e9bf0f0 not found: ID does not exist" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.173553 4829 scope.go:117] "RemoveContainer" containerID="c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc" Oct 02 09:05:25 crc kubenswrapper[4829]: E1002 09:05:25.173821 4829 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc\": container with ID starting with c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc not found: ID does not exist" containerID="c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.173844 4829 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc"} err="failed to get container status \"c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc\": rpc error: code = NotFound desc = could not find container \"c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc\": container with ID starting with c5e6ff597224d645f0c55000e94a18b687a62539214d20d717d4ee1aed268cdc not found: ID does not exist" Oct 02 09:05:25 crc kubenswrapper[4829]: I1002 09:05:25.480066 4829 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a98eac3-f47e-4b62-991d-9031531884cb" path="/var/lib/kubelet/pods/6a98eac3-f47e-4b62-991d-9031531884cb/volumes" Oct 02 09:05:30 crc kubenswrapper[4829]: I1002 09:05:30.461331 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:05:30 crc kubenswrapper[4829]: E1002 09:05:30.461881 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:05:44 crc kubenswrapper[4829]: I1002 09:05:44.461421 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:05:44 crc kubenswrapper[4829]: E1002 09:05:44.462016 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:05:56 crc kubenswrapper[4829]: I1002 09:05:56.461761 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:05:56 crc kubenswrapper[4829]: E1002 09:05:56.463078 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:06:10 crc kubenswrapper[4829]: I1002 09:06:10.461668 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:06:10 crc kubenswrapper[4829]: E1002 09:06:10.462435 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:06:22 crc kubenswrapper[4829]: I1002 09:06:22.460482 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:06:22 crc kubenswrapper[4829]: E1002 09:06:22.461219 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" Oct 02 09:06:36 crc kubenswrapper[4829]: I1002 09:06:36.461557 4829 scope.go:117] "RemoveContainer" containerID="638d2f3567f85f6e3025f92f77bb7845ecfcf12eb989f337c45c393abb9e5d5c" Oct 02 09:06:36 crc kubenswrapper[4829]: E1002 09:06:36.462434 4829 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8rvq4_openshift-machine-config-operator(60aa0a81-354e-46f1-ab0c-b1eb386974a6)\"" pod="openshift-machine-config-operator/machine-config-daemon-8rvq4" podUID="60aa0a81-354e-46f1-ab0c-b1eb386974a6" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067440256024456 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067440257017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067422757016524 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067422757015474 5ustar corecore